var/home/core/zuul-output/0000755000175000017500000000000015154265740014536 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015154271531015475 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000115312315154271454020264 0ustar corecore,sikubelet.log_o[;r)Br'o -n(!9t%Cs7}g/غIs,r.k9GfͅRJEڤ펯_ˎ6_o#oVݏKf핷ox[o8W56!Kޒ/h3_.gSeq5v(×_~^ǿq]n>߮}+ԏbś E^"Y^-Vۋz7wH׋0g"ŒGǯguz|ny;#)a "b BLc?^^4[ftlR%KF^j 8DΆgS^Kz۞_W#|`zIlp_@oEy5 fs&2x*g+W4m ɭiE߳Kfn!#Šgv cXk?`;'`&R7߿YKS'owHF6":=3Ȑ 3xҝd){Ts}cZ%BdARO#-o"D"ޮrFg4" 0ʡPBU[fi;dYu' IAgfPF:c0Ys66q tH6#.`$vlLH}ޭA㑝V0>|J\Pg\W#NqɌDSd1d9nT#Abn q1J# !8,$RNI? j!bE"o j/o\E`r"hA ós yi\_.!=A(%Ud,QwC}F][UVYE NQGn0Ƞɻ>.ww}(o./WY<͉#5O H 'wo6C9yg|O~ €' S[q?,!yq%a:y<\tunL h%$Ǥ].v y[W_` \r/Ɛ%aޗ' B.-^ mQYd'xP2ewEڊL|^ͣrZg7n͐AG%ʷr<>; 2W>h?y|(G>ClsXT(VIx$(J:&~CQpkۗgVKx*lJ3o|s`<՛=JPBUGߩnX#;4ٻO2{Fݫr~AreFj?wQC9yO|$UvވkZoIfzC|]|[>ӸUKҳt17ä$ ֈm maUNvS_$qrMY QOΨN!㞊;4U^Z/ QB?q3En.اeI"X#gZ+Xk?povR]8~깮$b@n3xh!|t{: CºC{ 8Ѿm[ ~z/9آs;DPsif39HoN λC?; H^-¸oZ( +"@@%'0MtW#:7erԮoQ#% H!PK)~U,jxQV^pΣ@Klb5)%L%7׷v] gv6دϾDD}c6  %T%St{kJ_O{*Z8Y CEO+'HqZY PTUJ2dic3w ?YQgpa` Z_0΁?kMPc_Ԝ*΄Bs`kmJ?t 53@հ1hr}=5t;at 9:I_|AאM'NO;uD,z҄R K&Nh c{A`?2ZҘ[a-0V&2D[d#L6l\Jk}8gf) afs'oIf'mf\>UxR ks J)'u4iLaNIc2qdNA&aLQVD R0*06V۽棬mpھ*V I{a 0Ҟҝ>Ϗ ,ȓw`Ȅ/2Zjǽ}W4D)3N*[kPF =trSE *b9ē7$ M_8.Ç"q ChCMAgSdL0#W+CUu"k"圀̲F9,,&h'ZJz4U\d +( 7EqڏuC+]CEF 8'9@OVvnNbm: X„RDXfיa }fqG*YƩ{P0K=( $hC=h2@M+ `@P4Re]1he}k|]eO,v^ȹ [=zX[tꆯI7c<ۃ'BdIc*Qqk&60XdGY!D ' @{!b4ִ s Exb 5dKߤKߒ'&YILұ4q6y{&G`%$8Tt ȥ#5vGVO2Қ;m#NS8}d0Q?zLV3\LuOx:,|$;rVauNjk-ؘPꐤ`FD'JɻXC&{>.}y7Z,).Y톯h7n%PAUË/_xY~7.w47mnjGgG{9_e552s4IG^ۃn󨔖I@[ tWv Fyw9J֥WmN^<.eܢMρ'JÖŢո%gQ=p2YaI"&ư%# yCùXz!bm5uAߙXC90뼯nNNXYt\oP@gOV ]cӰJ:^q';E=-dZB4']a.QO:#'6RE'E3 */HAYk|z|ءPQgOJӚ:ƞŵ׉5'{#ޢ1c qw zǽ0 2mK:ȔsGdurWMF*֢v|EC#{usSMiI S/jﴍ8wPVC P2EU:F4!ʢlQHZ9E CBU)Y(S8)c yO[E}Lc&ld\{ELO3芷AgX*;RgXGdCgX JgX2*Ъ3:O7ǭ3ږA :}d,ZByXϯ&Ksg3["66hŢFD&iQCFd4%h= z{tKmdߟ9i {A.:Mw~^`X\u6|6rcIF3b9O:j 2IN…D% YCUI}~;XI썋Fqil><UKkZ{iqi :íy˧FR1u)X9 f΁U ~5batx|ELU:T'Tស[G*ݧ ؽZK̡O6rLmȰ (T$ n#b@hpj:˾ojs)M/8`$:) X+ҧSaۥzw}^P1J%+P:Dsƫ%z; +g 0հc0E) 3jƯ?e|miȄ{g6R/wD_tՄ.F+HP'AE; J j"b~\b$BrW XWz<%fpG"m%6PGEH^*JL֗J)oEv[Ң߃x[䚒}0BOnYr猸p$nu?ݣ RF]NHw2k혿q}lrCy u)xF$Z83Ec罋}[εUX%}< ݻln"sv&{b%^AAoۺ(I#hKD:Bߩ#蘈f=9oN*.Ѓ M#JC1?tean`3-SHq$2[ĜSjXRx?}-m6Mw'yR3q㕐)HW'X1BEb $xd(21i)//_і/Cޮm0VKz>I; >d[5Z=4>5!!T@[4 1.x XF`,?Hh]b-#3J( &uz u8.00-(9ŽZcX Jٯ^蒋*k.\MA/Xp9VqNo}#ƓOފgv[r*hy| IϭR-$$m!-W'wTi:4F5^z3/[{1LK[2nM|[<\t=3^qOp4y}|B}yu}뚬"P.ԘBn방u<#< A Q(j%e1!gkqiP(-ʢ-b7$66|*f\#ߍp{8sx[o%}wS`ýͽ>^U_S1VF20:d T2$47mSl*#lzFP_3yb.63>NKnJۦ^4*rB쑓:5Ǧ٨C.1`mU]+y_:,eXX맻c5ޖSwe݊O4L)69 War)|VϟT;Cq%KK-*i ѩQٰ`DݎGu( 꿢\cXn }7Ҫa nG{Y bcWa?\34 P U!7 _* kTuwmUr%ԀjƮĀdU#^ۈӕ3ΊeBO`^}ܖj49lnAvoI "%\;OF& wctغBܮl##mϸ.6p5k0C5PdKB g:=G<$w 24 6e/!~߽f)Q UbshY5mseڠ5_m4(sgz1v&YN2姟d4"?oWNW݃yh~%DTt^W7q.@ L⃳662G,:* $: e~7[/P%F on~$dƹɥO"dޢt|BpYqc@P`ڄj҆anCѢMU sf`Yɇك]@Rɯ?ٽf? ntպ$ˣ>TDNIGW .Z#YmDvS|]F)5vSsiExţ=8#r&ᘡĩDȈ\d cRKw*#zJ9tT :<XK*ɤwoJarExfKB4t@y[6OO6qDfEz]1,ʹB֒H ֱw;SpM8hGG&ƫEJި_1N`Ac2 GP)"nD&D #-aGoz%<ѡh (jF9L`fMN]eʮ"3_q7:.rRGT;}:֪a$)gPSj0j3hLư/7:D-F۶c}87uixoxG+5EekV{:_d* |a%ĉUHSR0=>u)oQCC;^u'}8H0]+ES,n?UU{ x~ʓOy_>?/>l8MrHID2VSsMX^"NۯDc558c&'K0L /C5YDqNe~ض˸nErc֋@aw*r܀0 a {RQXV-/p:MP\<=<^越a/bz?ܓvjIg3MN4:]U]STa,@OKdٻҦ62L0ډ"ܺ_z9JNȯ=@oUI y4 A(" 뭗R==9!nKErHc1FYbQ F;v?ob-ڈFalG*rEX}HAP'Hҷ$qM9(AHx!AF 26qxCdP!NZgҽ9l*(H Žڒ;̼|%D Ɖ`Pj . ֈ,ixp`ttOKBDޙ''aLA2s0(G2E<I:xsB.ȼ*d42I:<ŋu#~us{dW<2~sQ37.&lOľu74c?MՏړ@ -N*CB=i3,qjGkUտu6k Cb8hs&sM@-=X(i7=@He%ISd$&iA|i MiʏݸT{r[j顒x.Ƞ"m@Hy_I )j|s#RGI!dTKL&4K>#stV \'xMgaSZNg8>e!^f%cYr]qs:"̊;isXa]d+"v=x7p.fZCg_Ys;pE&\U}ܫSh])qKYAـhhdEnU14&G * QIQs;rԩ.k83֖8Muqu_48dHܥlWW q>fu6+'}xu\Veelz`Zbym gp8펠ˋֆ:1IC8qٞ\vXçL ]X/r}7O}Wh,h ;RQ=]u00yiC۔I^3!?H6iUH:ô 4P$rT`%2Aq-֢׍qt=@x#~0)p# ы9'iri]ͪ/@繁qVGCڤr,DihB ,m 9 _$q3= A$IC"6g^4e`Xo(D*6"^eTh'4xpFڜe'fVQ7~'c L^ԯwIڣA.}H;Ë*׬=`^ 9]r鐃 -Dfi2|QwZk‹u^6DQ1&H凎c!n[mi3)WfsF:M"uҷs.1!뾧1%s,hQs|hx̗3%*v9(I;:'>uQ+v)vR/egBhAAdh]4H:nV$tHI98/)=mͭ ڐn}}~ק?g_6WĩDRc0]rY9'z .(jHI :{HG}HDN`h7@{jnE#[dz;n#y 9D*A$$"^)dVQ.(rO6ӟZw_Ȣaޒu'- ^_,G;U\cAAz7EtlLuoXuA}bT2H_*kIG?S(קjhg 5EF5uKkBYx-qCfqsn[?_r=V:х@mfVg,w}QJUtesYyt7Yr+"*DtO/o۷~|hw^5wE of7cꃱ.)7.u/}tPTGc 5tW> l/`I~>|灹mQ$>N |gZ ͜IH[RNOMTq~g d0/0Љ!yB.hH׽;}VLGp3I#8'xal&Ȑc$ d7?K6xAH1H#:f _tŒ^ hgiNas*@K{7tH*t쬆Ny497ͩ KVsVokwW&4*H'\ d$]Vmr달v9dB.bq:__xW|1=6 R3y^ E#LB ZaZd1,]ןkznxtK|v+`VZ3JϧC^|/{ś}r3 >6׳oƄ%VDSWn 0,qh! E-Z%ܹpU:&&fX+EǬ.ťqpNZܗÅxjsD|[,_4EqgMƒK6f/FXJRF>i XʽAQGwG%mgo 恤hˍJ_SgskwI\t`ﶘ080ƱQŀllKX@116fqo>NrU Ѣ9*|ãeeH7.z!<7zG4p9tV|̢T`˖E ;;,tTaIUle*$!>*mBA2,gJIn_kSz)JC]?X(OPJS3.}clݨ{e!MB,cB߮4af祋,1/_xq=fBRO0P'֫-kbM6Apw,GO2}MGK'#+սE^dˋf6Y bQEz}eҏnr_ ^O^W zw~Ȳ=sXअy{E|r,8ߙ!7up O`c Nc0%Ն R C%_ EV a"҅4 rT!Ddpč- .1j3Y QuH Ύ]n_2a62;VI/n|Lu>'&0&*m.)HzzBvU0h}~5[ Cnlnݔn"?W1;|ۇ=sx+@{.*+E2}`0C*0)QsU·8t^O+mXUq6UDöbV]/)fpOj4r@mo$Y#/7$&5m8TK&cF/5qX0mS换ohJ\Uz=b ZF~g&? 3 pnl58IS)`Ѓ' X'0ew9ƟE q& z?z^>%(ȊqmrdclSa iGߟ|ح d=UjZܴ n8=%Ml%>aU#-a`-$fhnqgTĔO5 ꐌSXzv  (OT19U}W3QvS}ll>,ŰAVG Y%.9Vnd8? ǫjU3[5U)OD:*Ϳ5U)=}/'χ"Q_5ILAٍkk7'VlWQVm0c:%UEh\1cazn2ͦ_EQP/2~ ʂK 0`~5"s#PCoT*,:[V4b=]N& B82^Wk9UHLPm))29ʱ UAcBC-|$M\^C!`}M^t+C~L@VC|J,A$SՉ8Y4c3z~$)*417l;V iэy(_άj]$9XN+/Sh]icc w0/ %IcT >toZ X]j?Xن,@aК(!6%DVga9@,'@—>+`ze" rd;L2k5x90YQOe%d1ؑHͮ4W\hx锎`qcU!m}xF^jc5?Ua,X nl^Cv>N*H0ӋG ^1v& ?_?Ho;~3)N.w7_|+qU_^?KhkQ\|ZUNϦ 1~u]ZTT걼/?Ykq{uu[^usqUKU\^|O-;:n;?ݸoCow-']swqߓ'~BmmK?|i [%ۿs& Z&el-leb)U)ժ,{]]Nk"?-KjLOg2hDz O\!3KXb^$gJ1_'9(ΟSYpX-ŮΣ8M,౬cxAX5xM"5XITd U$ZkNbr`C`kQ*t 0gb@񲱥-kc1VnR0s| fXD>hB֡#Э$+Jpᄟ,Cg:v xJ"}C[`ӨAFn5ʬL~ZPKoa_u` BI#DQdEB n+۰Z ?&s{ 6$U|C\l9"و5 |y ր|/#w0~$Di5~V*ss| &8֨O0v+jR,cepO}[ K\6F5vROq5}X-RÈlfXzqd5y ';)VKL]χ@b OIAG Lmc V >_ .z{IuCb񶍠z}(.>LC,HI~%.Op% 8& c*G3T\X)|*HN'1U0:VmoBlf`[ qB|APSO}E`́JPu#]8& 3Nr&}̼f*I vcJlm_m)d靕َ$4opjfљXwH+ƳC6җ2ct"*5Sct)eNq[ǪP@o`co ݎҀ9;Wi`Ε`p~mg˗%F|WPll{f_WJ|8(A ä>nlN"jN;/-R%~ {_'##AA:K`uihj$]pM+f># Ž^A12JQ̛3rFI_"*l§,̀+ å} .[c&ȩ f,M`,Lr5E} mX/)T/ϧ}V aU]<6Yn3rߍʇ)br\kF`XCnl}hSؗ΀ѩ آSOE(ȊD0Mޘ`MDN74Т C>F-}$VA:XB8fJWq&4flq6\X)ى<?Nwg>]dt.am.>!LcoJrKmqvcz܅EA:#u-sreHhl+ [cr:̓?W~cv>0Es%DƖ|2H\2+AaTaBˮ}L@dr_Wfc>$Ȃ1L-%΄{dɱL ;V[bp>!n&-աIJX)U'[؂2 KvQ8 I7D Oj1;V||`Us%Q)L͟^ CLu+."T#yrHhlكƼE-XO 2ze31cDBeWy!A/>l[whl"য&6 ' L,C{"./Ep.h>h8"oRs?]\McZϕ!1hKs`h0OG!L-wCln2&0Y'0=/T4G! H/ͺ|@bX)+{{T//seϙK$s`h nO/cE(Tcr!:8UL | 8 !:V^S7;G{u%[^g#pnX#&ߓTo_=2 sE0ڊSoq:+Ekٕ|dvv߭jjy6X î/ML-yz,ZlQ^oAnv -)xǺč--qcl @Վ*V߳`Ƅ(Nc}*ۈGmU`tiz5Ӄ~ &$yx `iKA"*1 [cplm_ugTvFW/ɚپi?~5I'wofڛT'uP_}9Gsո]qY7_G{Uw7עz(ע/BMI? m-8NJ\NABd!>_:h/AC;?_ϾqŽ6xMY(=ͯl~l8V0٩TAP*tLn$znlq&җ^s RTn|RKm;ԇZRCKTI|ʰ]Sj%-՞Az`sJLV/rʑ=IhM4fG'QIc(<ǹJi lc*a;Y޵3bRk'MmҜf "D։nʼnן$;vGmܨ}n<3Bΰdur$ Q,{ yTFY:dK UaRQE,$<I-CUaPu U! Rψ"/RlJrl!*J)b.y-k=`b_eo6,RGKmPr&Be~j%"O`R<**<@z; V3P6(p/?PUaQOǬ d_<m2Tu~tBS`&'i!.h8/8Oa =J. `0 vf2SRmlNI S7hYk=#0 1%F`^z@-A^W& ?yg!̐@{6ц+P3nS Ik!bܰj=HV<+^h9oЗ˲<9SɠM@^/[7]CZ 1"c g~O:Ȅt].@%n@=b"=!^(H'73Bn>7_~~S)XCٶ5C| ò8kׇ4tu{&P ]O3htY[AH~wMO榦kZ ^ Mp-`ݥ>9]&L|2XW:"n_T7\쵭ͣuՠ%[XPVHƀq\oW& ul:Bh`I<y^[w׾Xo>1<Ħ:q,"DH~}y cDzB Y~V鷺Zd2A"\5w8kyU3%-gQXM.z-®^X\OG%BeuΪ~,xfpOaI13rFOtRKF`ߥo^(JMikX2/HKdg>(pd(W-8"f oeg'~6 `y#, u|^[<kߦG( -fӑX P(HyE3/D:"X%` (V1*gYHtQ5. 5*s4ץb|#`.ͪ*/N.ςrR4k)IJU,{KQbZϖɚ*x]4a Y Igmw+}j$ ǣ8,DALlU % V囥_iZ7=`kz%l* [FV)+|w oȖOA%KݎZBk[Kxt4D6=5uwG퓃4{K SJ,<8AF9wYdЈR,.I;kUڣ`{ǝ.LjV\HϨNR [ 3A"b|wpM&8:qrx 4eU&U8D^ m>Y_,/g*rM],'ɋXGvweYHcxRX0gz2k؜Z !r)ѭGB=]q6z5[ Ο9x5k`ͭF_D~y<+qk=u4=|3?ܦݥy/?mj *ѧƝ ֛C0`R(@3fBai'YY5)r<L/{&"ϱ+|,RDk -망t5 >%Ĵ_Am@k;ʢΫyG`1t}w CEO8|8@A!5guz} _80xvvi>,VK؋(MG҆8ڛՉ/SLu?O'}0mۧ/|y_"z˒<ƉkKfih_F bMK6ۑ1Υ#px97kuZPȋAAѥYHS(k<T-iq-W8J_W3p*èD]T,ˮBHf@x@D!1uC$+ڢuxp߰E7solFZMCpL-]ÃI^dZ^'KyLw]EJYol$Qk'w_áq̧4]eyTo^[]G8ڮ=is7à VE6J ~5+{8,Lj:1TԲkKR-ߣy*jM\5X%LΎòŁm?_G"AZ%UR޾밎yTf]+W "f0 Eֽ-֔H$(fOgl4WXg@@%Y+rf8Z(=f wNp7^>H8 ,nm[ ((ML [Wkm."usjY-od>&p%Ryuռƪ W\~=†fLi.'EM?k<$~vTNYG}~*LdzR,KOoٲ 9*%ҹ![QKIʱ!%=!I_e6S"Xk4nnIL:be-qƶ.aY`&rLp]*XZF:љ)״!v`yV[5bGYۨҀ05cU9?Vؼ,׭MӬ+Fi[4=sdzGS!.47`u٣kRx؃ڒ&KJX#_Z _ia!">5A OqYVZ\zvԡ9 +TП]5µ$*Z"S [uּW 4U<Gޒ--b8`Q]E.ɘ_`Z떊^@}(ݦERGSgW}dˉnП,δjvtο"g7tu]؆]GUmPKDL!&V{"DT=G=EPշW;R9;yk@̯C.c[\= ԖMɐf b7JLoԥW)LC^5QkXUZsX΂ZhQ݆ƺPlMá*BSmY8^dø.g뾀s De]OvR!m g'noI([b&*!)WgXQgMֽmP(LFҾ*\V$w}ܯ˵/`JX@$o.VW2 AdCGIɂ\c]cZw9cH2=ϴ! CP7ԴBOn葐u |JaùLR#]yT p֚iVCPRo.ܒe(PmX*a3@p fe_٦ђW,kK[d7!m \R%?V`&xׇtUU<~w첅Dl#ur(z צlq -y$ْXcɩEx)+u3{Vz`'.ыѷUJ;mɫ"ĢN4%Uk]U}c~[lCqTHՃw&?ꪇQ>JY/> Hv]0qGΡ 9/+ce.76낆@frjM?NL71pnǷr| (,Æ9Ա[5By~f=|ډx@Bi7lq" O/'@hܿP-@; @; "GؽHܾ`?»d@FZR&fd!eo`. VcAd~;1N!NG0xnGS ^8hCJ84n!WԐw+‹%жr7a,KjNBC۵7 ϟ` da\QIM(O9wz8<4taZ'z4r":<ż# ώC.7%3#w5,e0b[˲,JdeydrN>8%d5 ..|8PXw"m2ed%ńX]3do`zݟe8aZԄ6X ieҸ輚H7AぁՕ f}?/+z,pk(S\OS^ j,QrL]i܉׉Ðavddݓcg3Y"UȎzŽ(El VhO !gnޏp٣K^ Zx#&p:}܋)* y$3tHO",?`u\xC>XփXw]oT}P?qAm\VkX<&dFާ60SJEf|ND*K_Ϲc3- ѧgyo|*K>%8<#wR7Ja #('ѾkZjժO]';gf$j:}: $Q_KqRIđU=?IMrm* 7DSQ| qK\?e<&~Ƽ .]%8=J{={- TMz/W0fWC`<sҡ\K,|4!ݱntU/N/셌Ŀ2 1T٨"C8, hք2yLUdjf5 lֵ|<\X'}(?P[7ZZ}CpY\X`K7y[H'&(J gw0˫y ^]9@&G^n:ϓeӴ(Np$br ncu /](p2[N,y6)6$p\ 3u 7lUG7ɛϦ""atlv ǡd~梸G Nzϋ=p?j.C%vko`ܦhz[{8l̍x| t Hlӣ(T[µ0r#h~imɬQG!Ѿ*}v8>wŬt5Vv+EF86܀mluG<.GZmEBQ˸*~ڊTc4h[P:/=RҀmANynI(mJw N(=P'tB׺ҧtN*ݷ[Nޱ;>]{hݹ.kRv N+;T'lB&:F'MBuw'=PwOBz;Nwޞz;7 w ߝP0B= w 4h@h;a{@(ow N(?P'|B&Fh'BWtcK ?M)PPL5׼u3.3͋@8FՃ]^%FDm@GH%5hD;a FeVV$.3J9zHTЏ|m>קn,{vqx?sw1<ΧY_BK$`VUF~ͧ(C̳gSy|qrKn\cbЫ̶5cyuϮ|,-)"`(d>mi?WYe^eiyI"[iGu4DU#fY"JSDCuj%\:f-)Oҋ&x݈gɜ c`("%"h} KZ*uk 5j' W҅ '2Z۷W6^/V=Xڲ"[^F@T[ao{@X-<9OQCL,Sʊɇ.C_|a2ɰ`pϚ(gY+.d`NlvM2Fr<<{Y^0eKOpˬn1K7CaTBE:5w-bg aMC%2N2@:ϧY20i@վ*3c/a~o Cl‚Yxx&Z)ZV&U5Z93`4jDvg=z鍄hs51"O_I+krhra 'D Įj5~LIȳ&Bj\7rsFA x2Nc;:TXRLࠞj`ncrlѦDٓ(J}6k[\G8Tf79zˊ|B! ӑuG<ߐN5v.<2qDqin"XiOn,>FZ4JȚ0H_]-MZ@q<Acf"!,GݝF #N W"|*XdĚ̉|fW?/7Z9:G| W殘*|fj``\ ؅ ao5kLxؾY0X݂ ?)vWZO4ੇGz mdE]-wWү=}gA(XE ~$-Xy8~TvDF1h_EEu6';y%:V?±|^e :pNQaґg?**Jaل^6O<'C4 )IN³ȗtm-O{WK9hj@x>[HˋH΁''LJo9;BrMo@h/SlI }&4 +AMUI&:BJ8fkDw":SݶьA;}Q~'̀-CgcH lV! i{nj $=)12C[SeIit.S 3C]0F5J̰x-bʓxfIKkjb R"P3naiټ̦oEq4&rEkwL%e@r,ݎ,/&.~D>nӷ oub?CĭZMS)e%*gV: e%[o\%(ǘ+\S̑l r ՙ0 C`1l8+VeQԣ+i ?>ϱpej0G'BIN!˞>ss)e92P84}H k])[B#ë+e G]?7N>8+VU,b ,rVv/UjR)4qbϬt` ў9Y6Z>49Ô)ҌQ3NaPc4\e%o꽦FʜS{ϙN^i=~*K*:CDΖ׳׭t\ WV uCmuZdM*cJS7TOjUʣLOL {a+VyQFpxAJ+h`)2ͧbm_.S=8;{فWnXh((O&џ. /Lvaa}H\ޚ(dv; (tkm4m}:>FmK%`Dbbʓ=%ԌviT~9,0jOQL'i @Au+̋V׈mG6ڰ[47N 4upvAa5IQL*r_0iol2pqCִ}pf}5#kҹ̢R5 ʂ[45?|ew, dmKhecma 5đ^Vn_A $S{e3jjM`H@4d(:D/o{6*@(I³fOkHgv ,  ]s8C͌H T_ΐ7@. &(VwO9lj?oV9~~y|7}Hy qLզ|逫b̏o?³]Ml7~{IS С/:I8y{wt;|:·ٔ3(μe+ׇ?ᓏjozjz.WbW \'vtD;iSɭG;M<#;g/>8x NַWb_Z)=?7m8Y NmT]ӏ/9TVl:`GuU%81W0؉{/eznk_|.(7P12;3LR3KQe/I,Gsr(pQ M3,x혎2/aJX9#zѼ#xvFw ཞ^xoO)ZYx 4b$>/:m (>ކH ZS$Ί`$i8g\?kVoEϒ#}ts!p#̂2uFűUB X\ RaHpl= ΏGjĔVi*(P79i布Ķ J:s=ϻWKpg9xxqzLE pˀ#ͪÚzؓ$0Ej&1Qu^Yz^ # o#uF#? tA)Li:WxYY4n F {&{V*$| ꐨ*!S*̀cg/B!yh\iZG1l{ D[R䲤i2p8)2ob;K%Í S2InXWݪapNuiVwG#p5F#1pp=?e+QgB4Z"X#|<8!*EUHU'mp Kkk/s\ay[,(GD+C3{8]QN됕E@\a}O Dn9ۄs 1؃fX<=8]=EuHT !m>".l6rl3 V"éӆU=%o-`$;aWHl58 ^쯛ZfYGVHpG^o{Ί s.ףh=_͚G3VMAE~-AFkLUcbAv0p<=Hr֍UJ֒lLYYشwY"ሑtjIro@cIgO ҿ.9Nu9;x}eEgj)VRՑ:ԫ!ySHC!O~Yv!jELO$8k^FF<8bbo\o3/C[aZW%D#j!q5:[ 5v 1xP"ו&#d1a]4l!Fp`I6qJ`GEhpKl"DH/h5x;&sL=-ꐒvmfnuepCxn){W]믓Qvᴃ_KpzWkI== 5|Aٚԙ>C'Ч.VW!1x{ʹW%} FNmF*'cKJF8\:] Hqyk1bbz{E:iFoi҉gӱD] >;z:1-xTew{*S5L唁Ś@Wc*JDKh$vo- v)-)pe9|/~y#)3{E[w6T#%1rKQh=ͩBgX2):UzM1 vw;ݰXy/م}ύd9 -=Ψo?^rlSq8('\f-2]+S-G? w~(}h%eeEH$xR)Q,'>+ E1*J:x&J˲q1S)h E0bcM7-U鹴}Fi'2 %^#9H҅5%WK$ɻ_"sT,9)Op%+.ltH#S8J).:7~KMNZM]{JJ84sbxaFovi_l ~33N ZhTx+x_5 K 3mXvG5\:y1L^h,`Y2f.B`Х@JHo A"OV8Sz31 Il%K+EH:1Zw8 'fȃ#5!SPW/uNQT)烋 KXa;/+bM9pBlj)k'K3}衖A< }٨~q~5ﷺ?Ĵ{ĝh?kH88p0pwp4 ^z[K,v!<%/@.wT4ñxjAK>VR*FY5J2ݹR΍U{f+#ܵI /a4گ7_(|Fnh#ûb6nv0Rv(ˤ$ .eafj&tA53y]g .}uQ`'YFҵGO10 ralByt0Q6TRo65&Z҂׌Q9 Ϊ.ypȞ2bF҉P9xq)6evݣ` S+υ@ FwJw DUCjXVa)m)mxxO7\Vlj }ؼ{8cP݇~I"rДucs+M}y6C P+bhg/KIsi9+OMniT9b#鼺H|ѼKy^ؿqQ/)0βo vtn:jZ0I*(x}Qs S%`?du7聭x8|ˢ.5z3ѻY1aϕ[;j)p1qNU 6&\ #c_'& ۫JI ⢼%:} ѿ_rV\ش%G*pt.d4,fhVa34Yzm$`k4,?AMžpP/M?wvYo{#_:~Qe9<^?AC`&Ib0x|)ٌj}L^:w$(JuQ()G@8SN>lfl80_?ސ88+fˆcjb3* ]DK*~lVG( x`F 'gl+׎t#ZxVc3b2{iLs ] yd9#iH:Ty,?(bM'0)j{`Zvd1%oO׆pʞ0eWwnp".9>2x桓{Y;DnNٚi99߮Wf ^ge4s #dFG-Eg:lv$8jqQe_DqspO*vP=HeK'&6y6Ip xK.MM7U QoQ)GyY43iE/Q\'w,m+baߪ5`~GeȉQD)WQ4gL8yϋzS:q %9%Gj^IƱ`D^:mIԬI}kiHfU_P ћ&D„]忿6sl( >vwꧪjEO Y\0~;ԙG^E< jA>lVY_q+8n,?CpeEK]ǭͦgώ1Z,c$9킀ty Gs;it`V9i=ٱ\%= tߎ}cgA$<E qT"h _XnyK\?ww1j91{Me]߬k0\ 3Ow/' 鴆MMU5]kzGMTEF-z 1q0T?9Uo9鼇WaF&174f;%i9t6OLeٝ+ii!ț\Ld-t|H2$䛳 ݜt~#0J3*gaX&hԨ{fʪE/R4_K- \/z=FƂ%z0<~&Kp }-p?AxݓλkD4qDYMi=Joyr|YLv")8 @QG0\Vt5z~{g=szT^] ^YÓοS` !?.{w'f\\tP:PM׌FUr<"C"d9ǂ1NGSN}-{#d/ hu òCȡtS/)o=@p@ܿ4i//YTF] mG൓\5Z eZr3W{n}37,E<$9ep dŤDeQ8=4-%Gn2:@-ʾ­ W/(S1dL5@K|'J+JY:ЬQ._08 ތMF?et^' fAӽXA7{ֻlLWΑ__^L-S JK:&Jv熜J¼>z%4< Ȃ[s_!8 pPun9E  5bY0\[+1kѵuK{45m5M#$ / ‡Q9 fAjM[5f7jzo\nr rit:k"jL(\6.rhJ器6Ŧi،ThBX#NQ{}3fwWe=.aөL`h FAJ>vN,u4y0CR7Wc4E7Q(h<\v0R:bF7EYecWƅQSUWU?C]*~L'8V`צ{aT`5kƄj-V=Sd~՘YZ = u [/痳faZWYXٻۢɚUIh!|  !Alɻp^K]7@onj 2޼+~} ԓMi漓.Diy~I}oHR@ѥ%:[R,6M؟wn٪ޝ>>5m= g ½M&ڷ"ֳ, +à z#(wRfo\QzT҄D}JfՐVPC}fY8TaN,w-/Œҽti()A T.h3Nxu/'%?N,ZPLJtfS3RjGza *d,##f3--kq-m{?iP;ȕkc{#' {)`y$>ȲVJZڝZ $Xe}x77{-?+%Kٖ`3aBvG\|9IZK}} <-9(Ga5CjYzQmquk.\"/^ϴ1M&B-5KXUnK)jdi[z?+жMk'8+;}sؕFfACA1?W4MU U߯6Me*K{A&0yg+1PQx[y Ud3_5ש/ԄOݼx/eAt1%m:yR6 KHWp)hKnMzP+$*E<{@jg /`cn]Dzȋ[!ޟCQ쿦\$l͹ d/NosB^Oo`BlmۯIßI-1 s9gݍA6; l ]3#7*/eB=MEn}e)+u%sz]jb%ߍƃ!XfځDRUwfK؊j'=BݏUُox:})_FҺη;:&<95cO?) dA@.qViTH$"P-KVih*J$ٴ<|ffV\F#b𭹈蒩Um:.  7.X uS vqpTx#!Λ*3踈w5>iOSof4Z'ǛaݎKwݲWO`\~4n=v*4}Щ2iضPNۆYT|Qv2F닭h7k)\êBbC9AjĿSW{f h^ȟ'|P*to* 9=5I}9AHsȀ]q/8d{w[O-*Cםwg!έez8 Eb1r,ISEi(2L7(0VL5vgmxHCcXX3iB.CĀ:o'{$f۵DU %diٻ6,Wfc$Xb֣`I2 DH-)9GYMN [zW;,qrEƔJd"g B4RZ6FeVm" 2lF8 yFmZYs"hWlZ1):p̛(N:'Ҋ͡N͡^3=uqds2lۼE;N#ВBr b FR:S[l:N#\"U0 @5TgG* R$Z%6@Hg*uq͓T 1\u&$Ii8CNS:6 'M ؏deJHAy̵kfky3bejy³8Kq4X2 LFm`&T`wW. š5Y+ڤMZ˭lQґƦ(&gD"rc9e%8:r$?0U穌(R d)7Yqb$P+XXbFFTA$D3D83"L38LMatS(kl\&ʖ 6j7nDV5*LPY)T&lh1h=н"H>IZ ىRc\CLԪC s!R F~R*InZXlZlZMSb1qG w 0F7+ߗKYT"$t":7p*hiO)+֣c]5 "MT 1D}XX"H+*1h 8`bKat#b`2BYP`F!2V_ =bbӦ a37oʵom4~eq2[~E]]--]= *^/BZ٘/_VjD(5麫W8<ao/r!z.\sdXLikpblCǡbҀ8"*SX$ J%֍CE4+-\9Ok14CnñSp1ypWwc̃kwT+8#| j!Rwp/ o}L}'v?i@ͫ< @D$O^ߟx<"?%ԏ(i\ٞ:(*xO]S\X Xߤ"s"e}q[8Nazu&׿ 7ejfmƠsw vW7}rBT_3ϟ>~)k4'tb?BR_ DR~/N 38}Ŕ9WrSzt.^=N5/nn@ݷӬR$IOvm?,3RށE7`}ՅgQCw#y}3)t/UxZNf:L6$]WL+sTP^߹4fWM|WݜW;qemq4D̲>3/GY֡dqT)rǍS]U7a{oE-XR6 5$]%P",Z[ {.+|qAh51&.tAjS R[Eh0.LϺwFG\c&mzdgMB`ƚyI(8$ֵQs7byjXR9G2!(h7 &#j3ѧmb,j`Rpݥ7мY\)zTkNT i5M <ՁJ*v&hPN>mo{am\R]5SMnIA(/QѬR7[!LK]è~xn\@C9@n,[-ٲZRs3 >y0 PӞ`` >? '|ppY4xZjoMH[ LoO-0%hf%v@H,ھǖG@MO%VW!s0]qZ֫ԓ@zstլՓǕB'%=\~Y꡸cƝ5eN#K\ӫ#3\U3`evTCzZ2Qjf}c(v>^R8mF”FXD!h^:?, ?'?\4s;4?(̓ψ(n4 yYOx ,.]qJx#|U½?văt\_it $(QW s/;2ɐ"Uh[EGv~{+ih4r?,R0[/auFY1m_ JәH?&)Y1LF8NY%*bTzJx`,1F:VDF&V ~9Ýbci>u&e==xYYTafndqGw!@j=?:\?52q)1$`+b?kެ萈Nj?pKn5m^Օ^'eQ&2QpbLOz#)ҾPo^^X>5M9zc=1߻kwu7 5!\n85N=8rm6`"^TQF"R*XZHc̀#"2Aԁx LI ؆G,?ZwPކ~9ʓM&C= ׇҬAAAۇ6}P/mE[4%&yܛiLOMc `zNieמOvءyɏ 6ͻ h~jg>%=#*%Nᖢ %`ZIKezHh_`VC+3mO[lmCmǍ+hW>8zVB t+$YtzǔМ=E̢,A<(܄@4m9-%I z =Tf=rVBKnZ3|!9M>oM5 JeWj$(ʸBe;R/[ =,ɮ׽$y=*pV)1du!5z 7K.rz2b )%y3cNS\|h"v$r_-#F֬u5MsAwGfVlc:?_^^70􇡼t=]-_rOV'ٜTݪj|-U$B S8$5sj\u[zV=}%,)31S+#}V@D*!;\rAD>)J>2DrtLKz89 RD&N3)z=PiX4R/K[\!^”LH$еÊ-`\/#Nf$R!-ރu佊M(5DYxJEz]*L4LԼz=@KþiS8>U2)E{WBakR1uT:|rPW:=H+AșLTX8R-]U0yGH׶C+r8-VpC8<.WQ{@WjS2Up(s<ȄHFH-P]/И>*cy E^4- Mo40 JF25e1[1hz%^)UHA{MyX4f--91`J4PF2%vWӱ Txg/n煺:\ᶮ~:kn/[sWCs2No럮5Y^؃w -DZжڱu :)>s󋫋;~'!>O}͓_W|OOvۭ;k?\77gܝ 7*><:A/H=v}ۮC+s;BU:E>kUΆ/~:__|Pa` JmW;LOĒ1Y'֎}{b5bM Zs X;yy90"!Uׁ (QakͺJ⽶'\GAߚk #W}ifd0t@>#}9*\XK;?I?;ԗY-=PiLR!9n׈٫hV AP+iQ6u I3\_/zEy5 jlLàҰF#tM1r_v!tVwݜuhRXoS=UX#T E"H|ʏ[T!gp0,hΓ^]$;ϵߝ~zU?̯#FA8E5國Mďdׯ'~<[mZ`BMhKb_5m1`8,>g#"09firUG6K3 0o~o~Z! p_*JgZG2TWšzhOvګ<SD|?bRlhdbSXG_É[%{Ll/Lh|JcIAX&PW1 JXbS/ƪ%!Xgd#}V4\[[y2 9ߡI2j0i vm.!<t,nt"N † KYbi,+֨ߨb@09 pd8;ˊj&%JYeyZQY1 AX'/ }IAr;Wˊ5MOZʊ/59`l4Z&TFVr\k}nO|&' \dfʆзeÉ s#x54TR`4$0JJ6ys7mIɾ>%Kz p2aHCBl9)ݕfZ$D4^9[aGB $vɇJ D4I/Ui 1O5#_j)5yOXFHmabd#a7.-e2: RF2Bb7۲eqzEG& $8`|DxbQq)`^ocHcJכ$ < *xzeyOU1*]uhtmߊZeAcUڶڪ6H[o>9٩ t\+3R@t:$vXpX5XBZU,kX}2IPIؖvtJQH)s1R4n;Y :zR}ݺ#C}܍8r s1AyUdj7}A"c֏őa^*.()'ꀸTZ7Q1/ik!N, k; AW B *i`UU&ԙ7aa?aH؃Ks5AV™@mq5/q50_vM(GADi'1y M3[AbVa,fk!wM*hAZ" @$efy5yT iMT$!d6H6 a3.>dʇ8@{W|2>d)!HqaZRhN D8m!h\S׮E|҇X1W2$*&BZ_s BC] +YʄSbZ3̰E $j6H>Bek T5xT,z-T]h ᨬewh<9)H>ubm@w&`i>|dI|l@AޫPlѹ (uWEJ3Kߏ)ÈI Iv3IEfkI>~wgͧrBJN"Glb˜ 0=kɛֹCM;5vhNPK"~&[׌ H("hU!NbFH_FF>:$F\)IV'y7`QQīLl|01u^]ȊH\a[Qe"0Fʚj\+I>V7o/Kq rV'S?Ε$XeՈ S&V4YBAoN3b8RHW ]mR2VB0`4 aӔ 92Xȡy?A10fIll!=ߓm%$C[z_&!~#)~>F>6#p#Uo 8/"`2r!L9u) ={#ptBgKB/bL`;0fWhS2$Dǻn;P~;KH9" H]J:/%%%=:ӑQ`8)اHz>JmISM?_\ V>"o_s3ﻫ_!?|//y!Um|,w0Ƌ<"_sw$[?|hW R䮉 ՚Vėne_T gs==zډ 5eEv(sWYx|ɆO/cƫH<$sEr9w\i>4C$/Zf=aϖݒp f$A3t$Eлa1A${.1(z>H6@XР/~b'x/wFcxӐdz5ն@e *ɞR&DlQ[ HA7oȤ2R.]mt<Ɣy؄!NT9'Mٵ'mQc6=edm2}0`mn^Ҧ@&ZW%dyubuv6yjx0oߝ[wr?s`=fx.. S_O_k/㥜tmo~y1cewR=XmqO٫sو`z < ? M An' QLJ[s uO_Ъ Ce蟮iBUWC?*E;ODqf|gx;}9C<6ә3]۶=d6{? z]oIrW"?#ȭn}b|CH3e10`ӜaOuuW_utN it2@}+C:* ZBy'oa ^ Z)2zCu.*MgO h&U;ei#@}{bPʿQyD5'ܝݮIpRjV7:Yoqzwn-ۡC2p %j0A o~؂4vEpƞ0'w&1pAˋ{۲E= T6@+{^0Ù .c{JA~^|1;i !;t?urZs,\6W'EXexG8SfG- qw/rNg=ٞXK+yoPκn!aT*>Fq i3v|>Esٟ|$؟0ߟ0DNeYX4T)W"f4;@j5!e𡱭K:oWGb}QK4 4MoIVZaD藙PTisɏ^7pRnC*`E4kB_{0j*1iv\n=MXjHg ,[ݧseO~B HR,eCnE&\Xk/=>=&P:cY;(mEC==,*i=~U{5Uy(}P%H[V3R_3)\[Gif@NPDsͼ Ezai䘒1DgKt)2rIշOG*˨mH6-: C,(Fiek99@$`TCbmG1N9SK?եg>T>Pp4__O'=gܝDoe&j+EQ,5 )d$퓤O7Iڀ%PQ ,/'bP/r 7yx8fEav_yGkv?b?4 B+Ց.uD\u00t<| yd Fsz\ p+v,vƓXmh|/j愪XW9BǓ/XH4~;3&Ŵlv/qp]&aIY6iXx0ht -qa5 .7Φғczy Zk}_F{xs|lxxjXTnFOPOnZKc4]|JˉOҊV=HZAa^VNQ5;vS[ȃb4[-*[fmf&w7+}dXK#f@ ELN>8w y`&bò^4-}/Ufn>f洸RrSu 5TY1:?w/OFy||Ey|u{񏛽WP`Y P[_*@TPR_9&i2=ή2uEV! =x; smg9y o(JS=58zZΤ5oT 1|,lffUM FԘ:Kd1Rc3"zTK)hL(-5J_@]شץCuSrKtTN6`$RAZ0zjYW̑6D~K-k6K@jCf7w/>0hvײÂf'i%KG.>.f-˞="Zb]vÂpi6Ki *SZ>\|7C3u{EbVXyV1>ɐN'R|vY輌sAwt~<&U9VSE`[[~?YIpszVq.OK9^oERyBooan۵# /8^\ˠjh+!K8e(/e'e/(ke.\J,o5ky1`Jtbǽe^լ_\)H//YƢM+:PILC=tM`aώÖzuVPp٢/v?,\pA)ֿ[N>RJ5 42e"LG+wT:'O)Z(w-hRw" Q &W) #-!F*oOze,O,e 8iyպjeLB`zzՐ8|1hjWޏ1di&sAAlֈQT7FzdЭd\ɔ&O^k)dѤ1_'RYݾ=ׄ(9!D#8x-W7gœ yΎ$%heʹ!l6nvE\dZa):7qgU3V)!gqyD٧A0jīr.f_ ^u{-U([\lgHGyQ늹-qnVA7yr`^ zA! ЫYv21a ݙ2<:I#Ty+'>Q JfeU:.|%W'7MGe69[6S7+7F70PELW8ŲKV$Bo W+/C'HiHVN7k˼I^S9qԂdR;# qc"k w} ƎmʦqtRRYw3 E*(t$ -["ejc.ORg'UZa %OiuJ.%X(3xCmJ}  KHLye,wVz=S[j WTl)g)C+%hY "S=LZXpl-$L]L-liQILf cfLQX38fAYrX`t I#JGwaESuzS.ӨazFj]6da aM,:sz1T@ߦ;DJrs6:@ %3bR.aʫLab}vhF%V4H0N`SƭV@ COa9dA ^HmTe207&dz4t 5ac?%A " - L #Q".DS r,3q+)$"00J@Wn`SbD-!h~BXWLL(ةH5FWkMV蒜êイ.P*{x0񖔧j($+pJs @V(a88ۧ^3ea$шk7,3IkxUF|LL6)ikmHa7~U? llb%EX{%EiʹxfHtuuթ{tۉ LoG nvcEcL@r|\1FlBu"3rbB }Fwv1&^8ZέZp %cc;-u"As^400x "ԁ#.m `M0dỒ&J-G W403)(yE n;Xr(^N k ʃZRF<@HE&rZ1ȼ*c|pLpta;X0G|x `"(Vg6<o UXTunLt#+QQgjbݑF9mE5Yr6j6p)`U1~v?>-u'sOY3/ / T+Qw`hv{sP6D. — 7Lܨ%ݨEe`{(%%(bDEk@QnSv%nk@{3ZMV78t-B̓xc MȒ,\\=vEQp Yds  @Q%UPբ`Qy(WyOc;*gɈ&BP DEL`k.lA|t9+s(FDh@e#o B)4I譈V ܭE AF~ %L0BM:̃v;ty1=! _ŋ1b:k2FQ'7P7jtLGs nC9'BwG!b1j0k̊Fj98)Ek dL]DX~}6( |˽ҌТ Ez <%2` rh*1lz(G+w{]r9Vr A2 5@rH҄OWLYPXBշ߬0q4 '~hu+BҔ*cI:*Pjr#. U &&~BE5FHmR5:Ms g=g@U p?vG- Y!~A\LV}w6ȶzGDHFh;Y};<]7FV2;;}'ˈ.>R'SDG!HL^X$\2cA]bYA_X+Cp/6ld?!ݻOY> bA^jTެ:?< .1+!y kƋW0J`[W NLris`zcOL%WaԻuqERa|Rm}}+[4ɫ뫣QcFE!ZA'RlERQOYNKc4x5Za"т |5Y2ƿ>b8?}jH׵Jv?@q*׳)ƒZr4e%F6֟>=B$Ta$`)yfw<5݁e큷ւȱCr]D%NֲaG XɧE;XBe1,n$` ]Vby,wo:}~ꬲ4(V~,Q6Xgn`IkOVZb,҇US?Fu~$`cFWmZo{ ` Fg AŲ r$w+OS/tZvw*4#1~,5|(X|~ Ï XՍw^鰏?hsH!q `ʌŲʋbYfHT5^,Yv$`Q)GaVX#'D-8x $HJ!7xI5>esq2|8^tw&N/.VNjGPߛO} |8W;?^;W}"Z{׳%LW-Pֻ|*Ts'Ϯʔ˫=ZGgG+L_>;>/\o5ΗK },gj1|ß }eY:Z[m6OnO9ߍl_ GgOV_~~ۋWw'nWh&~Öjӟo+)BChE4evρ.N;)'xf:t)ðݹ1Ѩ? e緧?x4ݹz fnYWq6;UO:>G^&>2/.f9.YDr7|}R\nN<^/9yy gahhT}zwj|1O\]peU~`l jb>Zқb\kc,6{$+Ժͯ')wB`] ܡ|0g0˖u2U')[1f%z?FcRQ|Z+b%Ǖ6nrLlY.q˭=R3A;w[mǦ 6ݨU-,D͖`ow'RUq^[l"i#)(adɪzr}_ *ńjQBr&Z"D"RbJEwED10wrktIېk'QL6b\m!w;UA I$Ԓ٥A&x)7vkPk,T2&#HcTF I݊-Šcd1V0-àS9./o@RkD$k$ը$:J-IRw KX̘=p:eZICsMՇSº, x3`ӭ^iֲeȲܒ;:"lP>&Fix*)5nyː)WL4VXZ\4b̖+kc^"g@xB!̠}E! |Re)v!dHOɄlƟ!eb67Btf*1D$GY0ۂډf-\ 2E%d1c퓒mSư+Bj} XHH I?Jawک-$^ 2RHj%6CB6 }Қ[WEצ@Jl${͋z#^f(фv%d)rcԈ%g3?Xuz$XղRȦB*Q7W*KI!oDj!_& F2U ѓ R(wtOp_]_ K**{ ܬdI.dP頡mR'#o{NfPհ֞b``Ѧ 4#z*KU\MRO5ER|ㄼLS6Aji\2gb@Pl=H!"Y!1k((+76idL g3 d'Xmam vM+&Y:Wj`o]J䐌1d aYo! h݂H,LhaꤧsQ1:\Э9 ,,ab0wr\ `k.AI!0ȱ>$D,)iN2ص.^J  U.LJQ!]`( UPۜ QzdQ dg q2i0&u3Jqw.Q>())Hb9*9k0́Bb|]vQtI #v9Z#b $ʰ[@B+Z`3&&2Й' \ (fĥ*"Ř-"Ik!)!>Gf0rA!F ްn}-l/B&6#fn>%‘v%k[0UJ&JWEB+UTe ,c*${;v :#.Vӭ*|" D&jZu`A>8F*] ܣy ^8!/:Y`P܎m@CÏ UdAX?Q iyG9 jZqة`M1~v?=n;ٮz0u4(p.{*2ɘ}6FAKzvCʑv_:"@ҵ=tW;nKg۩C׀R"T2G"<AnK )A:lKRJcEX0tt K *ąJlwgw!m頜 `soێաgG<N$"e#kq% t$@fjrZ%J/12@?A j!o<d*4\F?V|PƲJ 1#tӈTW(!c(ZWl-^ 1An#li PeVo &yVUC*"ؾ,et *50 PP_lVJ `J^ Qm4%w T2'jOQ2r8,Cɵa+3Hd~3 D…Pq7@\,rUnz\Q b!U@N6ϥ (be]0p$@BF rNޮXD䝫2BeG& ۥ^e_ޮzp7V'ʒ9q9'SR^i2׌s}8/~}D!bɃ[b̽RoSRl-e'OI粢-.Kz]Ew"89r}ȗ.oZjmkF?dYVM jqط_d=uV;:k⌌:LOob:ڹu-K70:GP~uQguQguQguQguQguQguQguQguQguQguQguQguQguQguQguQguQguQguQguQguQguQguQuP(e1w>F}ً1Ȃ:,, Ѩc0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ΋5ꀺ5'ufcA0ua=uC 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0΃A'{P0^RV5Nۗk߷~h+fY<\\Oii-~ˋ<ق2\`clK3N?,Q$ȣYMOn&`A` FgEƬ,X噀zfֈ),5,vpsYy.1 mf"r6bK8CО` ,*gVV:[z`Pkh=yc*mZ^}'uQ]<+[y5>_o.qE^I|GLbkt Xts\L4>X6y!9,EL@9Gγp$56;Cȴ3K»2 %f+sI]*ld# 6VlٚwS{!` [2]0ǹ7gs,xLไ3?l9ґ`9uVG/k#?7 24v&`}c?,"鹄Ǡ\(أ{` X0% lPLze‘`Yٹ{>`J>zX==F X)*f.1c{XRQ{7^+LK̲-2f6j>*X'XZi{T`+{ = @- C7[ԯ1}\h۽eiNMmdݥ ݥ +uRC }ovUWmZ:ÄbԤiwëbJړR*^`6U|D}Juw1|tqJime1Xޟ]]_ 'ips[즽Vޟz<\p_֧˫L{ɲSߜGφϧS\_o}w? 9QlUHU1[*!իK_7O7rӟoZ[rn&h룚 X-3.,io\bkp&`L"#+FG @}wQQ#)D+fud<W?ô&^LZ택 X2Ʊ X0Di&`9`X6sو%vO%#j.wKdzi_q R>d˝ǀpq?Mo<##&~8UɐAcgVZV& dV~ـuX>Ε7/ݽ YYprnC?߃(u+-V(75~ rEn%Iϝ):5#SWo7Ͳ_a^?k{mm|n^o|I|oߞsl#ȹ*Ry{=ק b׸77e{;7ysBãL: <rf.n6mWt~Z/,{G/K9 \*g7{7_'{_1W|mޜm.N1SV7{EP|Md66a|n|)|_X1B*H$(e<|mQ͆#?g|f~ 7~8~#"dw.ZCܛlk"t}үF[9i_AQ&f;7Y}c=o$ozE59 *Dֻ3$>=5Z֕"77xMNinV՘JZRU+CBʹR j׃>o=Iz:m ֗Z]h]Uu&XC@GEZ@CZ-ҷa@vvN:Uv#[j jXqf[K6ׂsM\Fk۫N=hSNZ1Z@.ˋTbwR>tRZY`RԂԴKM#$B"Sm{S'Lf.vWdN6{SZW\Qc/9%C ^~h`vљz^Vw\Jg *ͩ#mPw.XDe9w璟Pn'X'=h],!.CgTWuO݆V)YdwmY~,z? A<03 쇙͵DjDvEYM^$1>}ֽvשk Yg@r} p#(.Xȗ߂jxHyжq(2L<&q '`5dA ckh1J# IĄ"9-+d^0P>X5P/pt~;Xe1Fxq `a((AJv+7Vۀ8ڂ>U=(*`ɉ᫘w%,ܶ))e%)8 dxvƒv~]7բ!O‚.&jH4}6F84=w֌K zq:EjC$*f76>_GW1* Feu ((LFQF{q.:&oC@E@L1nA\{/6A9H#kFV8-5*8t((C \3vհGp gуhn e@P1"+YOE!Ai#: c;gQʐBPDYl?`9=!=eaB&0QIP@ěWT*-=+V^!-IρA@ f\M:)K `^ \:8Hcv~ƭӮ|ZVwkt>+k2 ZbeFx,L\AtB$8z`2o6:S>PpRHuܵ.0kH޵繨)8) k d@yJ=7?(5g80)Q^"o*(9Q%rcl>E#˥P4LT`=B*|Dz;`Q6Q1)2'V$bʅs urpl-H}lgZ" 2Zة T% EiQT5z{ֳT`-`\c#d&jâHYa=+NF"5R&vFo AX/;byPdҧ D2)I1e ou =>@ Mg TfQª8`Qi#2 :rfF8Jf:zA{!ֿ?HoQ%qt}k\J1,S/ku] ''}oᖎV:K7e.=ZN~ЛNsy%f<[<=>1O=H1?iYb=g :S2MVJZz(u0ZI^GcQuVG/ʍNM u=kB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&yBo$`g#\F#qr:@M uGjB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&iB&yB0C숄:jr4BX:.!67(1YP uP uP uP uP uP uP uP uP uP uP uP uP uP uP uP uP uP uP uP uP uP uP uPw΄w+~~w_Z0 7kp}/K^[QҨM"F%\Z{J] W`2h1pEM\X+BE W`{FclD p9c WV>zvE(7 pގ(\bf4 3pEh;ypnL=X;pEp?㱄+B}1H(o Wo<X ܿb]IkSh/zY<J޾>^͢0:ws%T(ag?>{d$?.<:r[m^?.*izY7[\O1uuNGjRw[ix=/ӴXn 0f&-N?|]݇Ev9ozmY uхrN몓~=ˍ%IzE8л!x qpX3p81lIOI|?J ܐH{Cww}n^yoVpvRUWtUҖbE_ \˲ftϟ}gyNI;o(|S; lJ{|~Z3 =x㶞E͝UTb:zNQh}9+e-e̱RdkY~ퟴMuRةMC 5'?Xh gC4v'_R4EL(/Ͽ|<n0ݬǿtJlЌuslrA(CR싦닒ey-O21F~y#-9> xUf=H=wp{@mj"jԈnY ~ͫ-W! Ljc3zPyy> |4b;^KKtk\X\絢>~|u[N)9jO(zOܓ/pɷUn ' 'JvHm_^]ᓇ x^8Y]grH~}_8~e>W2`坭bW.L.f=]]?>_6{dvK w`Akyџl:t/`66_ACFդF:(^}V>Ly8KNg~9Og''$I[Ui~qj v͋ V-Ɋqhy^\70]nf`/.\!~7?8-n:)}SXY"r}3n ]ᖗw.wo_ӿM~6C-xޅcNj>/{AKoGAvۦi3!y&Gwi=A~ T˺O|翇wj|QMa~Ay5$ŹrM mD6YtT&:ϲH67o|'l|'l|Iϭ|^'4?=1$nrֻujaѨ vtqLlWIGylJ0օI%,]Ev³PE7rV^9z8ϟ=&[(gs7gc w;Gy0U.T6u'.۰niW#ؽPAl S#ZR:YST) 9&-׸4.0nirڮ^R_%?՘,.9nC /JFش2Nػc`MVK.6S*pC +:]I#;r+92`s /G VWדՋRJTVR`.xՑ *d, ~\Ե >̇[#g r $ (A4""uCo%^&t܃t1#)]-MXjMʸ֒d}R">o;e޹A1tnP3Enr z횲!kU ;S`Di`&q|:.|tr^nӇvħyD4:PCOa1 WmduVWYH,|OJo`FH&ߞ4Q蝸j1#)\K'y/`颀p0ȉJL|1#+d^6ao[kt)')D K$ZǓ##W'+v:8'Z[M>61-7 {-ˁlqsek߶zpnEscHJDӉ$4Q:09gbف")lG:li؆GKfLEJՇAB|4*|Pr+xm{|5Nru o0AJzKUE v~{e!+rc<<‹5L;ϗf0+].?cG>ON 6Ҷpj6`JEh|2&|4r^[h'{$$YpU% <+ p\ާ_s(b^01De7]JcmG^+Q|7+|\)L̹@Ё 2ڴ njGF@|7PsEw֯{̢^^3Wi􈦆bLX=sI#q9VBV]bIo\.xN*тg ڔ˛W1}2V\~E>WfE$žϐdP iPoW.Y1}6턐x#X$Bi򍋅&ZoZ:##<10dE}bhst:iwԹsֆ:w ?7}OZlBB_-Սh%\#{mDĢQ099kD}p.W4@:Q ]Fi{I',M_?4󿓓#5;]һ?mK ؆r w0ь/׾n~[7 16M.Qh7~Ǜ 1V4Q$+t >xRf^y(: 43DB Œ+E`U0h>Ԙ!k0X]̞rٛE6^w{ݡt7}Sg]~1< q лXS%?Wx4YfS+&?{Čk E\i2YnWMxgU9sWQycu0op0"^nwr.&"fcC\㶳|'z:R,bܣ%5me@vj*1,H"$dnϣ/GG1,GGϫ5Ժ%[ $2qղl (:S)|mmE>SȌLӈ6u4a @ζIͽ _K$Z$-ZwKi4oqLqK˭#u*~oĂk|+?Z XJaCHRFpe|01N&ag`Lq#xK+C<ڪr⎿\(R ⎧)7뵺'=i?oP<{fv<0gɮݎ/}_>lk;mV(5H3܏Y;y?'Ew烢c=VgΒe$zx/%OA}+C~|I4{w*e/?8Y=n_ۇw.Gt3\@ Ǭ;w2~*bOg>{IRs!Ɩ$Z&hT>Y2£/@1/@K;y^&7ȀUN5VJuuyV-WU7Hx4$6`K%o򏶔4˜(Wa>.]\z@(QI3K;ϫ;m"ݮv[@9¤攃9O?3`q]Ź8{i'ܹ`.=GxT dw.nl&lG]rakM\?>o1-N^Mzd k  06|U)g]W%oMxqE:OhE]rBH2B.Dp$t2 E>oSQX9< MmP,a D2@\s_+ ܻo.?||5Dr5U}PO-Ux8 Rʞ2n['U#>?>~1İ#<|as0S"R[#<*qYO0Z_`o-R\*󴼮- šjl[S~K }|zC>SK޼o7}2G$p5 0n7GɞG"ƪ("!x5CA޳P-$خ}^xvTV1W? A9#dψa= 6WրW ^/M'tC=Qq[uk"iniLI!Nsi㌗'棓C>Sn2î -v\lzB&PJ2 1&{Z.׼W6bFVkU^N>~{;K#ɾ? A(s_;i87N.0sҽR2~WSZ8F~5 M:a;Nc6v:S[cMu> 0J$dT\Bt,,[`?qF0sE~ܶ4A5n%d1mpf<5Jy<QsZ[`Bb$f#$ Bl1,~ǧnPצsKҴ ?'})%ˠ 0C"yxD;5xP8ؐ+'B0$g*;}|~ߣZ{uBWٍ4uU^RU@ij҉O'cF>6Wn,{C4Z^G]v4-)>9ax9B=Mn~Yނ ߱ ie j-Ku31gSTbG^6.8zvś^]6᢫"7P><Lį2L.Y"i [! #;)E;R 1vGvӸǎtÞ5YkpV8|7Y@&C{"Cw9딣'bp"IOϐ%m Q狰}E+VHq]ػCjc}m9U*u%*un/~9m!xAޠc#hO)<'>䣝iJ8tV1ОHQd5e3?zxYI >4OW,S+C{A1"Ǒ9[I崱;W^Ѱ閧 &T1"kL]m}U1jm1KZXIЛ aՖQg3Ý"nB0ջ0>OpLrPSmk[KAdI<ͅcP37]}h>+Q%O#<-{ډ\Wք,Q7r,ނr Y o4V۝+>F8v"%9inlȈo5$:a`Fcٿ>DrHS^}!h֪F{ct=H (Xgf_b$3^ Ѡ4=e :ĶB 67%# W,:I :qZPud ƋSn}E+\-΄5ڽ21zeǔym(,xxrNR RFGZ7cxLΦ6قHi90=k6H9@pao${n(vcwlu,߯$]Zn.VU|T19ZA:dW|' zSbѝ,iLiBmYTߩ=ڿ#/My@>.&]XC;!赝0]XOc]YQY)ͨy*t&QT{7GV{AUx1(j]zh+Ċac˾XSͨNFo/z%+zSHpŕ7͋p(6Q!Ōx)c]lJ!}@hNG`H= #Ye17W;Z+= p,B;(chxZ.g̨%م_ɻy9|?4y M`ل^_4&/iLy^W oBd>?)h,BfާDnbyVMg"QNE@| HxF Pwo W \=?)F8;BF?cHP_$Z5ttӘD~)Dvax+x֑] E<֯I`^E&<[w,f|?^?T8T L\ບtw}{]L݄JR'Ew'qN/OZ,5t{4`$vtˢtۍʘљe )q6#!TpE+ф"ϲMhfmZIĉ:eiߊ>pH1މW@7xJxҬGLK>"Ziڂ`TOc .z/QIh͏Jtө%HmPβ&3D Zz±x⟳s{^a- ^.},]|D_{14ithcoc*M4ݼs7.%ĴR$)$R(5 iq6N('%nmjy\/57~#:}y_nUqmXOUӹ5)KZKri&PY:=j0Z @-PdU;[xgO` =G1`=p]'y.(WB3 +I&i(wRE Q;tߺ͠h)?r!Ȟ( X͖Ӛse%c0s%sGM`IB['2P/OKO=PCs\< 0d5I|z_mb[Imm-VlƕrQ"eQrwxuJ]y€孎=W8ʓGLF$Kʧ3Ar"\bLj7;͝ (, ̌~ʷx RE v2gYF,Jtسc9| 3:APdQ\/\9~9*vسc9|AcGՑ #ij43 vqA :fVJ FL;`FzŁ-DX7P8e㽯l 3(ힷoqCy+F z^`ZGc4$f(k*CVRB4Xp0B*$erdMUvfr)x"1bxtzj[tᜱ0#íbe=Q_;h4yR~ | nj>2i<'gc{ 4 FB`H :CZ,Ӡ'$ʈ9s>Χh4J?qGށ3gcmĈ@˅8 )r_ܳo:6Û,&rW5:kN nGwZFؙrT8)9߻zD"GT^ZRXr559㤬rNOČ_PCY׭ עAp8 h~x':.ML $&X,ewE`LQPP2,wn5j:_u&Fw:>, )\_)ߟXĘ yC9NĆЈ %#DBVz&?_m6:ۆzmƶӘD2dۋ& !U;b5E !UӘF2ؒ\҂G;]( $;e}ᥥԬ6Aӫ򪡂oCeŠSȺ32|HZMv?(X{썰P4_xn2|ˏP[{y0KK&E%$f4gVkuzk,MZMc L?)" RI8)IRN Vpb'{LǘtHT2v`lJKdBp!d !^#B019$ ӻXSsm9 wN2ڶ(#_@ h%$悻,Nr땵>}vdBLVwS #/c`m׻Kmڛvq9{TdmdQL%I`Jf4Kϼ}b@|5Sz*1؛$C06*3j m3Z(? T)2.3SYIy飽+ 0jfؽKb;"&\}[\046 |aC"ՆSon`x8xeᳳB,-se"Xn22T4JX|^ZU MZUiL hA5쳲qI&1:%scUQ 3o>-@ _no7/5tp4/*Anm/_AO"꡿xA?{'^Ҝ-Jm^_FN{pg,cƢ8`^z41j8 bs[T`y75KjK(d^=>d{Le)Y+)ԟF `%/`7oK{k}XY7MSQژ[bPI\of7y.=C"QȔ<50ca&D9LdYg/[W}tomJ.gr@O!:OK r:/+@h/.vp c/7>v[m[4,U ҟϏW]n]\ ,fLݽ.զD[q~ihv:òX peoƹԬ/?VȜH*۳\C߸%Ց̽/3wY2!?86uY P'seLΩy.9)RPNs$ gJ(Y}Ot?"+'M XGYEP4V|Zla>^=/gӃk8=/պKr^yۻ xwww8\t0UYk}{}ƭlO^ r>֛OOx{W {2];r7kI_-$;/M>Smq]p5fxq?.i-g]‹ݼfcޅzSx`흥Q2vkS`->wI%=k0s|lJOcS`9 y 7ϑ,tWR1o)A͈ު۴XI\.ۢG ALym kj \WK|]'ۧ vΈ\+>YmӨ{ PI7Fmg=d\*fJe+%qeR)7`w 2U9A ]?g*+/]/Ү7}kq)QJ$1ԥQ _Eaq 8WdFYuƉ.eŒ0.J*[E}Mq4$--%L`頱Rt#!Tj/u  ?Ibabh <4&Vz;gZwmF3_ndeӃW[^"57d]ܬ b4dx q<8vHo0a!.7?  fFWoy$mOŢ 7*p@^ݩ7Cώ' $?;skp9 f|}> ?}3,J ̏XwZ7>h4-nI>2Ng R"J[l]kC`BJN lbzw,p-27ߞ;ٔZ'/;eHs W-bpP6HH./F Ц#&7HSh$9tA62=bBu`cxZ,HmmKTC:Y\ÃMz k0WB lHlyРe1\ `ĩs~ZJI0WQ0) S<T@EQD!`8S0sJ//$ee]H] z䳉]4y7j:)0ZFeCRX AvTV%Zd<sU񞨲P)E+g5ҕ&1Eԋo`Y"XA|8 Q-dvpn[)I2{dǰ5tWx% ZӤhދB0gRƜKɄ2HȉTT7^Bo-eJqzZFap2}|!pZ`DikcNB322:.Q*}$eh-c;,2T(mkiNU\ܡ\m :fD)D}wM\.1^cɍs$,jG2m`(6 nD&p.Wa'uu19>YPwKnoO<6 aĽwV#G jp6k%g̙yI^QzQ0Pҍtk!#ur8C@8d"-զQr8)SpPكr\p&i-%!}sC&"&DvAMv^]v^J쮃bJZgי7ǐݸoO>E %A5`ߵXvGX4,b)G'YOi¿'hq`|"u㓸֍Uԍ?9V^`WZ˴&ym X25/7iF8|i5_v k*5 ?}>{FZAO;禔Js| ɧPyl_dYrұiǧۧ|IF=2ʂ#il@ݳO/0>e׼C/(-N_3^HZB y%蘘oz[XC\Kqax y)AGDO@r1J3鮷GFapdmz{?W8 dD 0m00*Qg˭P$I( 6v\lj`rc9q2& kn0c`VVF"GDh w퉴 iяh07!!I#C<{wRFe[f=>#/)ij^e~ J3t¸q9 <֊i(A Q1+ic"t{a@2Z&z9eMm[XF`B*%h%kaXjL0ﺸ!)"N c#08,)yulX 8t#08 #[1Z7gj9N3r!*%=>l( NJ5锿R՜\$ 1*y>$H!R ǚw-08:P&jAsY5l`O`wڿyLº > e!rl\@ԟ^; 'y`z@}֨N"7|Z̽|{s|_E)2;/t iߋZrλ`hK6M7$ʖ\"F5;~.%D1TN"Ka+y\lLM>Ư΁A Ⓦyk!]_8#XȲ§wpw7 |=3%ZբϺ-kp!Xq|\;CR\" C#L_ "5E,:bmW~-08RɞUg"<$Uߤ 5q|*\*Z^?4 āRXblkazUPB~cy%JR.{dGq(9 k0læ4AŽA:ψŎcp2 uP&*\ IhuT56$5Qsӕ \ˇ=J l+@Debϻ. J\SBdӗ M4lsayg \)|HpbM Y; ;XYi~3*3q3#)GFap8V$f> %xw;`oaQ0lМ)i4j3FxHr&tTsVôŋAaǝEDg=`Rq}c|n41-gY5X"7sbpoͯBu{?zL/"a* &fI'=2 iϲ֤P3W;yI+{E'M*#08Mf=hI0\sW$U+^W!)" .ϥEOJ.@O(a&4t`\H ABp®e`|qzF򤣰òEq@Qln 8\N*(eBU/9 :!n@ғx=#:s꫷筪\k%o U}I, .AYw{2Ev,es6[`*-P9h~2'%GFY]Et>R7,?X\`3inb1VP,!^g`=-팑xqFi}%B#7y@aNH>TT ܭ0U뙽 s+M0Q9lRosX3T8"'ָ9f҆Yr if'>G=D,hLJ1cɍs_eG] .)4 *r ^&1DzdӤkTd=*ZUWIl( ~p` \*톸Iѹ{!CM#Z5*$yQz=z4=O iWI*<㷆_2yHJXހUNuwȲ"O'qTZj\݌gx߲@y'S lc&EH&v ^E]6${IHV 0޵wrRlZR㾅v D=,Ob#QvBݳb;/ sӢT>*e倥 Dml'elybЎ7U]{){ߔzV.c"879oRcqe"Ҕ-M)mPƽC]a~8-*q:<}8Z(XV#GPn M.0 N00r2WN-q> el"W\3^*-eߺ?n)rS q;~lLJq<43ӿH!;9N?Yg)бw/",40I#w IJ{nIB/PI*p7q4Ԛy^Sj97I;7Ii;*)Y5֚@.('$ l;Lp Q =PS^c!|,v۽۰+%f0zypU 7*6kQϿifvټ߈{ww:/~>y#|&oUϻe?L=0qw5kji_vti0Ћ/a %&]}Ix<_d-O.?C~tʹ!|gmH ns`mpg?ž&L  gH"5MGm23g?xڗ;s=ds xo\Qax[,'b%/,&vhn>lE˴{پ/[$?l4o^BWDꢁ-ߖS4KW:? c5 .%?f:>[K<|Z5!kMĒAq /;*-%Q>P[겥Fo[tq"Jq{ճq"蛯Sod4fQ?,㍳Pc_O]6L|;/_  | X)VuRӍkY-ZGG(k @],.ggS 'txF<]?=Z2#BU搈NXaeN0v<}SvoO6n| VShդ7xjcyt6 3W[ &tU04x HX542UД"j;%r+I{AO)g2CYw9cJ!i$WL: Slb!0pu^)FQ),U]Crb}JL[([&0;e{ƹݺ-o>mym<")I?^#_ wmrY)uSLE%޵kխ緞ִ6b|״Rj0՚VǏ86Ы5FY)!),-%`[Q=o'˓g'?eU~i?aw: S0'^/g0t;= [ve<,@kd3(ŸD8\vUuyr'_RxP' qc(!L95۔M>oF6Ln:ȃwW,ޗ4䊮Qgu8T4WoV6e5zouMo;?|\_7UusedNiX|.g-17+VaN{$J&ѯ* }!A5*u酫Oy?( ؟P Jɐ4-{>;Iu?0 \{1AkYi[R*LE>,}o6NmFay2>>[3$C{kuq9 x-;HNR [foUި̭w10sB6Xɶ90210J9ؘM/>ਨȗւ.1$B"HxA(}`dx';nucmh׎կ'Ҍqι~XgJMutfVo{p&I1P;,l܀=032մv4k! v7) fa,Sp TtZIؖAe$ǑvdRY-;B0hKЁ9@qfյǾg ݂"'t/`8 DUTH%=$@VJ s1_ZK|4"˥`)ӥ/#wtni/vxBr jnx5<1VʰC{Ƙ >Aw;9fXU>@ݸ\^1ٱxE<|C͞0|+ȬS.ܿ4qN/wJ8wוqutxF/"[ b%UNi X0#"ilAY+T FH0]ʠuYR,W:K0eNW5є)&XJYܼ g'r2>tK݀]i$%qgQlH`J;LKJ Jϔ*8*C sqD L(?$ˎ`O#Oɗ"U1W]\El~ dݵ's+/ܽUDXwZa͕7..iDUjF)G;ƥw r>s9gAj^0wRun,}Uqrt0*2XiHU<daJQTQA!_ u߸X\miQx Gƿo: N+!G?lhntdz zC]zj@w~[*?mNXH}wkB`%2vSfy,e;EzWoMW6݉$'!^:bJn1|qؓ?p(!QIj]zbu?$NlJ5|n<%Τ0|_ϷTޠ|[|~Kof-YtJMG{ygKGcZW31bZuLL&G 뽓 j#1DQ<{lIщ}9=_N:g[ڌNSnw9nbUgz48uwG7PHnB^]ٍmf‰̓jpl?]+=no}^K:ׇtfs9]Zϐ"g+7e(`+oHn"GoxMKJk2H!>>Ւ~% {4]r4[&~eĒ/pR6]ɛ.+A\YBn( {dX_)Wmc>}ʘ{l|{٨a_@nv_~9M$ArZOmg=Ҝ\椀ȑDy4^`#195iݜl'e7D(yLaMd<]z4#>'b {2/'1`rLE+"x:mFtz.`K%}2W"SON17W!,z=y7tZs? KȮy2WO"@d`'qᢘ.{a`?;Ė{+g#|YA|d<zvrviggQy R >Dِ H|Bw ^Ip~9Rg'+$6]p z' 20f<;MJB r(&-s2fa&w!Dnu {Ox5M|B I3Oq=[͈CVpscJ-ħ@A<{M#e)qA! JbP@)k%ԗq2(cS ^~>[^ r m}L%U J-=*&>B⍖OC| sLB!^mRaM| 3s<Υj82[u +W=:! NlLIX gd7r[YR ^PJy~YLf'>TlPR@Z(g3ʨ}CA,K$C|,k;Oiq7)r&dXQ *+S d!^go RpB*b@%DcBzOxA@r_j])k]EL "%e=/B!>B٦0hnS $``v0oC~-X2ѫ9J&U͈ TqQ[S >2xA:=_M߷&+qAN*hHT=]q VxfɪU?^Nq{~6L-r~+CTP G q&x5$qıL #B^JxC|ijXRzM| s}K`5-90cCSD raQPQ-c?`KDY0W_0سOxɲ4-~ Yʅ=:!Jɀ\rVN}tT#gJ9n,!>B5bBuVB |m#Q} +LDngNh@hU뀈IYG|s|I X,%yo,t' LxZ,%#,! l.v^R\!MB#)rϙىwLS@42XK-y,(UQU}{)r/$ͽg9JO!HMVtBA L n +Cr"0WX!o1g$+h*r9*w %)4 @x!c!S@AakB3%GYw}:!. 1%5Aڀ*0Ղ@qΧ@A<Ϯq11Xrp(5!Tb_i0$&8LJD 24du)rVJ%FUqi$*@`yYMp8Q1#Gs{4+ıjoۣ_Qbrs_J~ۓ^j<Kޑ.ɠe!-.N^c[.Fۉ/ۢjN@N0MM}(p6ZT#2Vϳm,6X\m%z)9zWP֬˦EѝԽ 3WA~7suҒ|(ިRHU_ _ *0^}^;YEJ@ޮxuǫWX.l,j>,GOR4ϯ +HqxXo4RH;Ӻw{ތrO~~'5m/F3%p/ǫq^o$kld_v˝rSoZ_8/>o+q8/~(H^4qr37ٻ/]IZi[.-X~ iwmY_!eT! "ؙ̎3^N0E$eoU7&$[b7ɦAd٬WWru/G6L.@ >ɵB]9B@hVH  zgFKi[CJ!)-b>`~y⃸qnӘrEj w}؟vi?  ,9t `dr튪<d#7{K_Àwn}7oay#+BrBF].r9{6):#\wܫOԊ2|0s6BӇ1:ڋFr;@Yz(c3x<յ<˫Xn.n<v??B4X,/W+4)ۿ_,俷aK P ՠz#;vjը *|N_F{x.m)L^,4Jd&)Fٰk,^bsﵦ2`"|P!$䃒|PJAI>(0A!4$L|PJAI>(%$zΝj,!S|["(9AFX\s+^^^nXb2⟙̓g A977>ߗͳ̸?ʼ払Hkw@!K#G,NsC!3淼zs0P#cߎz # XaY ig~z-;LiTI1p=VSQJlN(>B%r&n:hl҃ cL*볘a]yk!<ݰ KOꚥy%\km&Tkqhe> Z,HRT:T:լN*լN5S J=./4ifOJgQ0F2SyTBmܙ~/Q`lJqnc}c6¼kvwST677 yRڤդ:mCЦ:mC{@ \SrN^z+?ɏ*2ׁg =݌wO:LiQ ڌA8.2e gH o{d"[YO|kQs5.Lj^rnb&^q0ۻgeNW=?F:vF̨ 35|"Qnǣ/_=G)-z,#э|gB2f;UT%ZFεmb*׉6~#<3J7)#,leh&O gHzH hASZ~a`Rޟ7o{LAx '>#o{S0x)~dNW_G-ns| y 7UW #zylڽ!2<%xFF{rpI#g%?|#]Թh';Nf@.  ]o7AM(X1aFTdAcV `{;sy^K5~˟㸣9 r$ʍ~gHSxƄV.c6$dAws~yuU{lmDhvb |4)m3:܉ǻv!~I1LX΄B=$[o|xi3_f@~~|I9 'h\ynPDc(_+/_W E?_$EGѨ7G/b<툃 WKh }w_!JI ۅt&_5oz+SW\X#6ƌ|k 9Q{UtJ4-<_zfMW=;d܋`k[Em a D!A/#ʕ[g~ ]b-,Lv!m qaʹ:b ҌlCkm[_I鱆k) Fmit&8~;TlLu:3!,  },csyi{T=k炠nߌ&1$hET6P S]*rWK[]Rj61_mH TzWV<>DCcdۄ\Nܨz,p ŇaQL%3Cx< 3iKkGAQܙ_RZψ2T#!Hs']=ZE;Uc[۬ﵡeu]Lp|V0-¢ٮ>&AQh]?τ+INn"N:-h2nhx wf<4+MY GFѕ,uBoEgrJuXmc=Y *.s$hM#_?򫖦_jYi#Z Jlۘ6+gd7Do9Ay=Xd#yyYˈYj< cy҂)]tc¡O.V6l3qHc3uW8}%6q ^M3[H{Y*=ǢE?7yR˸ɖ0NXo+g:yyxgvK}p3aYGGbp$c~1|D̿ 2̓ "7vG:W rUf)LD7Ԩ?%9xް2q(ɻ"b(c"tꜝ**O})e׺ |t[N:M4&)L6 AmmU<[儷5->H:jLr淉0z zOOD6!<{=x{Ma9,@->լ{uאؓqB*Şrѐ[rGSoMձ ^rw*t5Oy}+JZI3dSCo#V}i]-R`x»G\{6ɻlʁrJ-; qaFŒ䥸^{:,q{uS<$oey|nk? :Uh<^)F-tjFi-!!"C-T>fWj5bMܽGK 肓\~E|2;D@2 YW/^;728Is)@՜aX7`QqZnk (4C659lGK1fS: :?4kLpoDM a=(3ʭY(, +)Q>pa#v.C⮚ܵQ[q2ú;fґd@;Z`O#|T5(̸aD3b9d8 MȀ50Ɂtʞyʸ2ZPHL )$x`&a!aX8kJ[O-0:T%X31) P%J8Ė##i 99v˜gXqHL474c.-%NK&uZ`Ew;Pbd|Y3&U9FC;ݍ(eB ɦہoTv ֙EfS!N"h2Seٳ%79T{z`m邱ATP!4h+$@=. RZjȭCJ h=z ܗhѳ[?OEyV|sz l.е FQQ}6FG׺"r.|e[_~XRA8?t8OV]~}k(mAP%LVlԎct&֣4A#A*(n:SbSBJ^gP9ƱV\ O`YXA[ N'c(uc%&y$C vPX Jy .\.'y)/mmmZ3(trFBn$sB&a(Ao ᝂ<ð L hL[)c:N\5+ho}DI);R]oFW=wb/Y ݝd7'ɞLWMR%AKdjzqbn6Uտ"帳<<[1iA( ʹ7A2(#HԙQɼ (d%k^UA9M_=W[6 <j虲#n= *b;"(/QPWM, Bʹw>tv{ViSU[wTgW\"w]^XNGo0+a˭u} ݑK` ,olÑxNv8cshܙ^v.;:wD :\$0GoGu1qU>7؋7z!)n?#=h‚'n}<0M#v .pGOMbu:kHOnOj7+O͉[n{kFjPW+z7hWH-v?~5r))7^C|*D1yNqj's~\aDٛrN^H19@/e(klU7I BkɁm#Bu~A`] !84gXJ Mf`h67Tf aU/C-´%J{LѪE<)e` -2HI)m .rpjxp`ÃUAyõ b㑛?С̉:4_SKӴd9UhB(+9oqAtn"m=H]7/7!Y H#1VLUOz~u0gQ)v]|w[,s_He("YmsJDC:Vpa0vdž»|V𚿴lvSײ3\-(vx׎|0t֏{VK펡Wv=cZUVOrԞ:t\4.QnN1;stbgnvj=xmphkOndyxۻ/πQ [W c{׾.K;!vɶ2j[jS.5@ _"./_U>o'X ]|bmym(_){x!Â,ļ. L! 8"clXlpQ|s(ys'.ajFރ9y+k`#mfYЙ)E13Y|x &B%Ogx{]\ū*z Eb g Bue VyNW-vF3~Oiu5xW^*M ; 3FjD|' {}+kvu4ޏG`dzS&co+0Z$wXj^}0CgFq?6RI}iy 4L`4PH#$CT~ȁgQ}Z`¬}}0]x_>}0K)XZf Y/)9+xRG_bPL807P_csw\;o `cp"N#n, G+Pl#v5R3VBg~AW@'L<ƥ\|LJ4o#{qcn//w|Wj=hޒI70{Lg 1W >0NFw~&H%P,e*qs_|y0ʣB$jektWɧ{*U^Mԁioq~ZQ  3* ac)6˃ėԲZ݊N70f:aS N1zƤbC1 b2ϔv!,TY^놞8 g761QWTə*9S%gL3UrJTə*9S%gL3Ur3+9CTTKà.b[@q(D{CN4ɣy䇮 &%jC {d&R"dsjT9(XA tYja$HPD%ZNr׮UKF@-;h)M~ީ< ?y&0UϚ.FUϛZ~, {aUU8U-z!EKkWogZ|`ɜˌzM2C8ʱtmӗ+B;F"r\pV2X0J7^a1JqHZ Ȑp>XRD%VH9!f2X0`F4En=tvM"5!P߹:v%/r|-Ǧ\~ @T|nWv=E<7Vjjðq/'N(YHn1߈ m4C:X孂s]p)v^#;18Yߦt̑1\oLתk B\2X6D)|)2_?Gx 8BL~skf@ 72u|#uC!W3(f8054" Unang gʹ9j. ԉ|5Uՠ,"Ѱ}ՠ4U d-Mw9K͐[w\?9uj!0_G|u0 kƹ0% 0obBX AE 5POKj5C_ttbrlǬ,:_DzX]/K#!@9jE*̔!P.rD15ƎoKǡ/:H:}3f;;SjL.Gw'NRc ;nQ%t ǧD}FF\x!~" ez2zBRsvڶe[/\_k\\0&⸔.AV22idT 8D#Ĩڹ,IjRۍCQl)`ubw/dl@$Pat91rj8u+'xTS[!۰+ai=# md )Z(ci9pA-x^h݆=I$K:vF:$s.(8QT'(Q3VR=Gxi- 6DnC6$-%MK֥Z:P"ٳ^6xv "< W_^Ș4z(1v3mVDLi)3Jؙ[ "yF% vթy3eGz,0T8wE Q^&HK%ZfՉەZ5HgҼojqLe<ȡA5żvmW峏鞍ݗxs1o\wt"v7B̃IV.}ig^|8K&Y`,ĕaNFw~U[Rg,e*qxe,x42 6WN\eUFtwF;.ʛ#kt~y݂ ӌ/{sEz@`O"Sp y~M`0NC:$ eP ePV6|t4 jRhAt^,:˻ea|օ˫ `ъ{ے)'TKR`<2AZ,HDR \ǣ4l[~P(jKSeB]HED,=QIvlz_'ɦ,؉z RGrCGn"Wdl#]Jj 5(9}ott"5R_~ӟxx%7Flܩf)gu!Sޅԧ#v?{WƑ O3;nC qGvه]N #P7  &Md9dJ:++2k:Ж\Ayшj@!{>t@r9惠 vo\NQʻ Ɏ7r ަCI`],Mws%ݦCٮA.{kdkݥWqw2_ԼA;syre{ {$b],JgڈugKw;B-<7g>i;J(7c?[Nn6|0 ?Gi5 »7}56 ?GOxF7@zjNۇo`|Xm hKfu$Ȅΐ֢e@r WWoMVh vR'0}ӛI8\,?}0j)`9JV[3Ћ2(Pԁ,i+ܢWe.'P': *a?IC珮V~w[|poanz{[]*~U:^7WZf2V%|{U],ſ<($RofAwaMї[g0! P쒎?FoPT32@cmuDmV-cn\32W@@lUI:s%UltT2cߕ&ku~x?--J,W2],H8%I Rei81IGNPwc.=FB;,Pxw_h|˄J?ɪXj`Z^67?Wjuⵙ$؇r HUʳd!Ei!]Zj&RWPn3ɨg(mP3|ȘCX  }/׮{aXi~^c*pØ!Q`CT`HRG9kFr-P/p3k`IaLwSb12!K kAuTa TK\fDύ} @cNc y,q2(0 Bh% 0\rq+w/A/-9fH!сG`cZF1'9rrmpᙃOG4@9t$ԓo'Q6jzLMlk>vztJ^3K~wVڻv07MSUd'Ti#~E a(3 3hQ9b0XJBԀDL"ƜOFYܒFiyl =wM˨<'c_u:gg8G*sv"3\+˥r\2)LʥƲdɥr\(e.5Ke̥ƲRcX.KRcX.5KRcX.5KRc}RcERcX.5jZve|CS|#sx=mJ(m7NL%c0\. C`IJ/\YRp5mbRc(%LPqxU.L}P}g=g W#EiA ,\ v>e#4PM-A)XQGFQElG&A z kn,IR]hzQܖ҃YՎE4޳ {u ܱmtnb:8*u4OS(n8uq?PEC"+(Iн+Xg_=.ñ5\yQ( 3f4j|sWL#,8ZI}Nܻjnb}#JHvR|6 +OVmugw$'ô3W+|hs`܅\(H;!&zG=)xcu@ @{ɡ]Y[l3:ҒDGr\._X(@,d k,)PV8di@u!:_PFOFQ1G EmJёaYqGkZ )$du Z3N[ ?4F=#mnv 1&ϩn S;KĹMԲ/ٝs5R_ hcy8s\))f?ϯϣux3v AHu2LwZ D佖豉hj4BZ"Z3[K gQ|mLRe34t1[Ʈo Uֆ2<ž6$vj4=͛ge?ҽYֈIZ!(QWM?QM_iZ;I"Czn1Gd[y'^֛;-/n#͢ړ!L!so zj2Ca^}EvXhT9~u*/@gy8 B J-&>D`2O}0QIP^0| BIЃ}~_z3~?C;t_&|zQm`y+?.Y;=_w˵w?x* {7ռUM z89'VtSzWpB/?w|;9[ӉAgNd_0B/PK%EsZNy%64EIЬK C ls ƛeɒ"I,.J '-1S8cDFs$}E斋wdXD:(zF( J>*<%jkNDQPPFAI-]-՛/wTnZpHmw'*4aL"}̴d60JnM/zN0'pH\<`SZxSjd-l )hj<O)$lpG A.nǻmjDH] 3;I>#סa4}7u@1ߚxq$\j=b2/~=Fx375CoG_n7V#qt3/T{l (hN ]*5MB__{v S f|3(OS^'Ҷ݌{q2Xl%r^x4/dvՊGzBQ C:mLBL|r)~yk{(N0'l%rl%j5>uGHdB׍-瀨{7 + ͗n=gmogd?uV1*r.ze&z٫k8e>\r,pU0pօGǕZUGK߿NhTjPrhɾAK4fg82=Gf<76fQk$F6)kiBVod1a(@1W|2MC*MGQIirzPN3(K,#5ӻ\йJd-{VUy*V`Ў!M얪#v?ir}햓° M)&SN~7gncB;͆.5ayi1}dbPOntyfh%@ }e1EosfS`{7x ,P>Q JPQn )8f Gẃ/c2l~%VgVlU˺ P?uDem6v\B0Ngޓ$+_<aCӭ,iHAFx)*J$U!b1+̈8DIɬR\L!F(מy9 :zN8LX/:Lx hB~dI/B؆Q(fJM:nž+G}p@YyQfx0պ9ٚ8=DA:ұw;!]T̃X-TaTĞHǪ[L l:QYo ;}ΓΧՃ%>|9\΁8agW V͹l@~U3_6s3^䫵GU?p.l_ٓ]<6Dv^.@F 0R;z&4L!?9"'\el9,gMx7 ]SAnN_YFՋczԲ묬cF'WNy2;ďD?P{LMpNL'P PC 7\5ry,WMy@)R^5 D&9 fSNO<  x%8&"9Ad4i<|\/j sdQ47_/֍ޘ4!r^T 1979~ZmM}ar?#2o =ޮ9w+ ywӛ7Ǽ)5L-hKJe+ቨ,0S%5e`DԨMԉ)VJ-II"zsx Nf%UFʠcKJ~(Lΐ,%ŕ%q9/KQ(a=:DB$-QQGJ De.)hfؤZhsȪQiɅ+*uPCXAt` 84@ q|$\rG(d' 2A=ɓ^=6P< ? U8CH6 Q8ϧR]n㸲U ~]D+ӚWW9XEt9a4"Z@V9ΈG1*Aw6H\ 4DiH=;" Za.Ak[϶ ZB#u@g/_̑zg/t Jd.hTU>V‘P\Ef _ *]S(G0iT *x5A<Ei= :/n_gB鐳Oϱa mjb >-S =^jdSu)PF7n Z" dՕA7[5qS5z]ng+l![ո_6IJl$eVPaHe{ik=bJfTyy9*ON e8Eؐ*F)x&ּc1 7)V߃aL.+ 6Tdoa<r`|/kM)(4XZE*Rqʁ3Ub޴@ipB$tL)|6 vW,{; Aq-7ѯ h#d7Y}RP-wB/1YRqY6/\*^%Ucrp θKpIl[t撥PT",7)`z"eTVCXZaɐ.w܂O%7i!8yȍﭚ%tdr&pJh`ua]iĸд ^5l{N0;A`o6k茀+}yj⏝k޸226%4BFQ6XgvU*]|JP>uԊzUW[F^].A+@-wS2,ӛ7;|5D_ _ܰlU" {ytHp.yYBlwZex#3I3P.'#fg륝ţ={= ^f f #m{fZwx!&T8Sn~ 9[&.@"11gM@/mx6-_Wgreq&(E]NU\bAjKf7qS8%{\Ǘe7;в3"O7c2K?2RB(}<2E:El.'ᑉ7= Ki7_P/wq"]x4>6Zsn m烟ѫ{ޭ1n/b>Twmְ XtmB=pm501@QNrb1r=>`7WŸ9рO•GnUT;цʬѢr *oiD.JhCdް-{"g-JYM\By SDr< 8}ۿm>]5UQ!w\- RΔ2{TtG*{;f쑂} sz8,}Ե3gQ(z14c:F}2>(]i譿aĽI 3'99&NYo\2 bDubXj%€OI,W]N<$zg5&S2EbW1E iL4ZE'%HeEPZq;.pQ[JP S=' F,HW@g"!ysދ]n:<؈ZIfUk i;pzftY d]d200BP<$W`EG 1)l gA`˓UǨͳS)( :𷉡 L"hIc.b@AT#XيaBS,O$m=IJIb"ńtT e }LMR !z Sh,DMjqе<[/;6lñ}܊ekqj+"7D>-tWFݥn=n/Uu߸/ܰY0HIJ;d IMM<4B (Af8z1-xkUk޼˰!!eX:R7Jf 7Wu@뷋_GvkHlw)V) #<72}6fGF$Sj: H@@}pO/ruw-XQ YeGPڈ㿣9WBGmqa(ET_rϕoSFeRR-<)hT^n Ss ,JY88P6Ց#(o.7x[яm6@ajv G(_*\4:{Hg{HMT#$_WπIbWiqEOh 84U(a5 Pa_3$(4D`Q!,b1ԣhIRRkFU1I4FOC+ī NV(ВL"p[3"b&ӬM^mەΏ~Taw`IX8;kj${Xb&4gP+#X?[)7+<TΣj.\4U6N+nG5]c//}QkJ^|M*Q 5|=mѩLx ݲ!\UMfH-g?MO/+@Vl6֮j~nO|#{ m`=`2SE;(0vEB(aje ==)[B<f!J;8/B~cF˕+Pt% ^ȻgFwѹ!Z[WK¾Pnڃ]ok@͸+x피.o l85DFƎr<>[;^Jc~p>+shFG.|1RDGLjFb)Y ã .^DJE8 Xp #_,vY|)ּcueMzÅ/yq:rS=c!342Xd'ca ݀$h7n]톒%22Ql{ &p.5vGjamԆ([:2Bh&~ECi^{Dj tM0Hv,[Q)S5hT>Ѝ epx6Ȝɹr<8RB /L48L@YZ B C!2/Wx<tujBl:΍7]h( (J\ by yzjݦKۡ@L(]HƈO/Wuc^ ૻ68nYԌU횱Wi藹UdWZSJ$1{S{S,`3A b!KZ .[D=Ƃ.9ժ9URHcDr OfFpH#X"z[&!Ǔm"8x&.xv=vT?~%e%!auG* s6J/M^78/ٻmeW<^iyNr }DQr/Zv/ӿRv[%2#Tb__8PR p_/X%k$^n aA$9JÎѝX6-׿~?gnEtC?[ ޼8RYaWH8j=r/֏$QVckXJ n"=UӀE6w}8g^K PA Ֆl<[dSk*T7i}EKUfw8^QuQ4IQ T;++#9+(YUud+eUg>' ի8Gy)ke<]f䘼)}ER21Gzww=5l}㲋#&("Dt^曓:(T$sy& Jʄ/|u]Rbim@;01xk:MW97?o 20 ӕ CmLNC/z1CiE̽v, A!1gtzd :Jw+8nܸrCoo>K b-=!NP^v5{Zmni<(U؋MZ 2E;܆utfDf:Zy^7[p7N٭~`.gTBIp_Ι8YYw1i{܎rkh bjw $iEGOH,0%"N sHb; j4I7r);&́Yν!)ܾ;SjC`u|(YǥӖ[d-~BEWI`$<6C&G{Q%#1jZ ,GVJTFPɌQT\=Qދen6дR$8z{+xnڝZkk<€qc0$.mw.0NO_o>&$i;yO c~Jl ng%{S "tn-ǿ?Oe-FP;%Ӻ7YLN # ) Ƒ( X?t}ةBkot*SԼؔ)-l鸷;RMrcWչ*g PQJ 5@2:X)f;HAeFF\҇Gw6#f @KTJth-`f5lhe.1֓aŐJǬֳ8\;FĮbОN`C?-_vct=Eɵd(.*L tkOS.˶r|}U5cy/CybUr ba i=:z5V+7SYqBqpN8K¯{-'qz,7$a1#[ '֏\lyc@4!C[7&U0V{ g:@?nS]m%LVغjL49myӣ(5Lvtν U DS|mЁG$k:1j*R>+-Bg7|Qrs20;~&CBF(! AlJXlH5ŭ N=<6†3$pANLF H,7CECtc|O" KB*?]j>P@)+mf1eERխV?%1EWW*FY`dWmU3/\JI._(b(!Z@TrCÉL(HRA^)zjRo#>JQUP7JR³bw9KCCՏ2C#OMҀMTH p((ZxM9*I׿~v8RD)d>0 FFxRX8Q 3lzx z|CsWflx;ch6dpխ[\uK SsY[>إBױ@O ;S^i|栤Oʄ>1`"3)ʁcWd6%bs@FJ`2QS88bc I/9t5r1\S>/M&܆^)AО)mvXa2^ G_P0VRra_#@Xމ`y ]ѝ8gT51&]ݏ[&ƆbK?k[nW&M ʞqXd˧es|F!9֘S!4`Z5Vu-ZGռ+?`u+w asսʠ<_0d:ٗk|Gs%;wn¾sx.N߉oF)/-o|2ĖMjyA< Tklz)ibW%RP@T ;z7NGUm}Zފ2 zS|ӬNk [@,S:uCµ<8aZm]sZxF%OKSҶє\né|sZ@=x#uKTX(L(/h(Xce3{gYXhJ~7c ̇t2Ÿ)r\|pry" tJ+֢(Qz-r!h1 8nr˱#Z=ֳ'ދ͒DR]1v1pԥ|=|'ڱƐ5,7@{C ]~Dĩy,dBWc?gase8軎5ec9s sNSMjbz1}hsIv"e$mT{heXbb S2{w/y>njR񹽏9Z~䃖'l{U>+T5]xmW(E.Q[(yQ$eZll']lEv[k.ERvzz9|yίO(.Gx+6,!bA_&*YsB1X*;9 }Edh)v#2='.6/z yd@1 w  OV~Xs@gX ּx{/ة;Un)x4u`'ScaRŚ|-q b .L{{F NBM`|&NI滲́/nGctoVhڇ !.4K׎%O2Чlv6٬a$erzk8MX?t35,WXT]+.졘*}w JX~JtPkyCކ' ނ[gl|IJC4W&*W<㻲bCC̬0U5Oab@f&w .oqu&X>kT1Ɓvv%nz Mڗ r9e~:pw4n7;~Ҙjcg@̬S )R[W$i F6%b8yJOՉ^Av+wC2-۟1A3*/GWXx pN;@H2BX ^Kd$: yC5"VP6y9&U"`PHxn,_|Twi"ۯ~LDp=]0y0S}s4Pyx 1,?}rQ6PN YPǜØesVQ/a!AE&HԺJjV/V`Ry1 9Ba 5񤡴4X=Wc+q$Gί"4.G Ø/  T:ZʺzJS$S*IX`w; #87f7_l ƹ\-L(ѸoyUiX(a$jautaE:*ùØ)*[y=D2EeB0'lΜwAqYk'Z@%RBxtɋYK( o\@(<àםp5VE ,ZL] |*G~6kn(u&ϝg tj. jw{ʨ߳Ӵ1 =\()gUe9z*{*ʦ)Bb<ջ]]v\T0ܬ1z*t`ogמ_=FSB**đp߷ }$sK5Oet'{Gnf؟ }*>aV6ΊeL}P'|)WS'3ݹϙ~ndijN93By;$nʷ=u.A9Fh^Ȭ$N6(?5o ds‘G}ޅ ÿ(Dj ^H.n7㊶yq.|ڽfy҅WTh͑_]q'ڣn0_Džh`$/(nj/!+0\`eixEM:KPMp4P.{Ʊ7*tڋ~v#a)M-67ҩ SIl80E-E?6}Е#> Ds哓J:Bʙ`#N{\ShʽSmyشZZd*!$ J;͚n NeeB_m[ƚ 7Qq޺znOМC:V֯NZ32hu3NP66#q`+~{ezz]{qbF ߶匯72tGny 7lYaҡ\5v(0wVI#$ EɌVs7AxF"{eB>T P ՗\j'waUȻUKN/iev8?)_$%O' +D+e:X%yyo=P' 19v՞PɅLEqTs/ |:o,loxr8@]Z@]mwHPm坙 ՞X*,jO.8Dj+M%3XA-k=[G1Emڱ4`U|`eg Vv\2ǢGbܠڲe`XonۋT\≇!$ i*ɿ9!X-d?" nd4i<-:Sl;QNDesWw[((rH rޞSǮNzs Fz~\E\cߦ 5r} ꪮKA J_[g[nm@ZpMH#,$CGQv6| F] \'K{Q%]vo* ɳ~+y/HI"X+3a62W)ulf 8A߳ȶfUt}HC\|]{=v^zT́~ѾHw`3,ŤorV?u@u!]w Klo DFI,-y Ҷ%Ue>"JK5-Է4u@V-v.s*6Bs\[^u%;> q,c=ivcs{D "!wJӃ$ = M LeڤZ- &8"c! V@~mKD4hV)I I;.vWG]$aAbq j#|lwB-<]ȇ G8˧@LyG;d4)YrtR5JBnN@ -[̸=`u24c=Xa7\@&vbn4:< +3%-7#IK~. )`7`Z{ТQ>0C@U|Lp/Lsi==KucpK D?f'QRSq|p*C6I~XlIZt#ܩp%p 8AM$縧[rB ;.W`BJVq@ho8ⲜٶYhѯԁvb-O|?Z %lj@*[qsHx`gI"rAd"j<cN)QeOz5Ch*!Oe֟[ZAV~E1qe\fE qJROe4'!x" 4s4!ؿηXS @p6`^6k ?.R;}5L L-=ˆS ;"~(Դ|z@Z믣4fyFQi<$9a}4Xǒt0ꌆIr-)!E{߁ M%nA^Mz iY69UY>#L0&m [Ҁ~U*;<<&(hZ)Y v'G<`,eQ5_(|WsI AB5ϳh"1.axX2&Y_tcࢵfj(9}[3ܯ9*F:(s b4 ?'{MNdFJ$2b8YYeӰ4N(%b/w  - , EFpQ21 S#C ,okԲ-ԅH..iXвe|AУI3z"./eD(I"n @a@-A~X@C23@0{Yk/ps5hU,y* L1T]sWs!Y_ O@t[ 7hIW^_Fqj,ggaF8qN*Ý;zޅC#1]`zM@$~y+kȣ {K4"\pom[KM/cg+ ʠS\w\58:3Lvc5G_N@>IoE\oM֤oM^| F(c3eL; 18֬3{6cN?W5;}Uw^yB{$"P`-r '%7Tg.8Q!Jh?e]߀ X4}%y_`4.7pw ;C_)عAkXjώ%[5[dX"+ .M:8Q19q 0Qh,2^Ck7Xu!',R"uq\A^ŦgO2Q z&Cp *{^#'3RP@?gT1=É\vADąNLZSgPBkmȎ e[5[g0PT*A甧Su*+˔$@D.y,qn e\r2;)j&اc5 [5 zr8iE&,+ Pqrt\Ѥ,ӂL4'(g%X=u\ :.~XpxO`޿SS+P\Ѣ:G$D|GFXݚ[S~wknwu 2ΘQ:HHf*ih N$ 4F )EVs[+Uׅn-z5*0C[w=Z]XAVج(XN ___z9Xg_|bUOFuyy俙4Po~~8Ф=.jEOMJRh 'h75TJف)Nk45w|I,<9dV{A!=% p:-{עxή/X1uء-j 38>shV4&[f Ҁ Ipy{рtm0~^N]{ۮ/ᴛ>z_oVGZ#p17Z:5 \bI еTHD'hi%aGVc%XQ.mG H0Jt33Ҫ=U Pwem,~Š~{!%ZQDҲ =HMV2Mc:Y,`yC_~j@W^5>o|*k+8U[XQb kɇ9{&4n:Ŭ\ q縋m>A_bƬ wk8B2^T x5 A0&;|TcHw?۩khele9PR~SɜM*޿oxkfk9|oţhф7%&79-J@cu.PiG7aT[@7;/EO%&}S9[-@hfosnc~qKi9k*-K(w#KCT8IvPO^^ǽU %kA\'ur ip$<Dw[b6 NR1aם4jF;礐^4y=> բ9\;`~Zd 8f QA_"E\*X4RMn/&* l@XHYzb~$1R2G4IjA; 5$_x7# $Lh$]2!8/9ht@o޷XP~/s#s^6ٗ4!DT $KLkhbz>/DžRZIXB8J :T3 %C깿 v+ gA8H*,i1׳9r.ESJmI|'.j{*nv)Q ֒&*i"){ wDc1<uqscՕԀ N׎ ugamAg_?ھI)3t'U ckqDt"=I%4BSFcp22%' +]V9sE{\A-< nKg_>{ ocOgmy+nKjeѼ2 d*"MGJGQ}9}hv ԧpeGg3;vǠw ?|H @w`M݅o;t κOL&>KQHrX2!dVk-7G~Kt> O<47GX`o1ƃtdI ٓII[]dMzVCCФ<*WkApi{RJ0;9r!ܓr':2岨Z=rp `^OK 甴NªpgT0R#:KO716y#--A܏%mL.k(3$$Yߣ&a)I{OrgA ƃqӀt9+)ɪI-i<lj;fXx% "ؠGNh%e8yc-SoK*`96n<78z?E%]bÃC%&XQ tV~z5ox5+Rrb+1و$.[;n02DŔK*~r8@pPk]`l$0j\c, G+@cu z M^z 1{52b'!=44sqxkE\ KP$]2ʆP^x5:Jz>N槒CO￞~Bppٿ/9b ͧ5N~:^nz5חLPW9tjϧ''fwgi~ۿ/ ?*vyՌw|^OnGqv8yŢx'g;d ƹwdYWN<*Nv^o$ppX 4=3)5 >^Z6ݰy5[uܧJjw AW\o>ǀثc,-'g2߉FdXHQkW&Ͽ[PH=jЪO Xt;A^T Wb!P^:*eQ*vq$]ch}יMjE !䣾r] ZhM7\e)#C&-sg 2J51^3s+ps~!艪jMt6עbS9Ef㒆5~ҐRɂs 'E?-4\ӣA.59Ko.`^*' ]9}WB{DEmM>mGƌ(;hQ ~h&O r[" h`%R xmRTR^5]L*ؕP,8AT}+rERppss,Q``4O+Ȇ۫\XU2P@L/Fո| Zܖeڐ|&$~K1(*hLW"=x*Arj@ɹV";H]fJ!4 CV_g K 0]WUJ6[05{{&Rn><]H.aϖ!9Bj,b/{yEFqF!ƨZi%g2ޅќ{#B:]W}:FҝdRf P#.͙"Xi2lǶ٨.YODBhNk34 ) .e划S\8iAjP6WE ,EE5%jFI " |88h^mqn\ЧY%G6N[V FjW}ډM}CjRҲ\sY0EQ6HmQs픔뢬ϜĥC*?'Y弡3/$2d`2aZt7upCaFJ f05iNjN3McEO[U~L& :az1+/0gY(#lP8iMChBc ?]q0*/JVȋpD2ُ-Ppo'"9IjB`(z)AI4h ${3A>5xb'7qv?^q4TB @}өj/^;Cjg "u꠺e\;@äORwDBNFhSm]`b$O'#EibkIB-1m;c~3LN#c{4d?A}iE_ ?dB]I-phn>kZu(xkq4{`2l0HO*谾.`K;nbdS`'0w&*ʧaka75`z%[`2OxOE*ei?`xp d2Ne~0aܕxlG6DGIk4 ;dCܧ+O;P%'X+7û78D BT{lOz~Q4-8¦e5۫&f au!'pYP|-!u9^]/b+'cWxśiQly5 ;w5R)nEYˣL+ӏ2X=W&LC/J2xZ+imhh._G%|!g_- 3gHWF&+6NL劆.Y$`,L dJ| CSL-Ų ]~HASPE-xѲ-CGb040CUO)FbxSe͊97,=_/~>M>r>tamrTspˊrwćN%[-7o{^-ɹ[&{b#7YKH2F$/4l XOFdyIENY9(_/x ܱ|5Z RpWX, Tpj9kKp.+.#UhqXw}*℡׉Tzz>BDž21H*!;I촊+XIN%#bDA"¾s=kQZB`ܵ07(g_co !NO.ٵ>fp:2"AGGF *lwcdςGmk'\C6;͞N [+8hcbUJb6V *x.8ௌʲm!!IdњQ; yA7zTE"A) * Љ/ru|F &+Z;< Vw5nr##V &4C5 Q1VZefdƃIuxFnUB]SAyFଖt+)iJ8wyt%Z3kdLe b^+A)]_atKyԽ$^XxHNfNl𤒪kEvE( 2vu6-[M>Ng}Ǔd12_X/)< K\^^TڵbN'HpJ⃍F 8+-KLp`| 8py~ ׷K\c-t A<)8]TbU"(egtO"[HR&pvzH )Z·e`ӥR yq*Y)g#\ زL`54u-21JTӟ>$k3$&Mq؛{($դ8}ӗWҫqtѡXJ;jE%ޤN,z9 l(fǰ97^s3\G%̸Thyɇ8U;ݑ,'^?&Wv|%ȃ ׆VeӁYAJ(E%(/yaUÖv;< h`xmZ>Zhf=iL}+[^zvbd}i|,f >YCHGqq J3˳YI=hWM(SZ"qr,@VۋOV$hn{`^6E0Ow .I^~MtV N8 Y鹚*6&hsA"\L-XP]x"4Kw6v A>AHplF%@Hٱߛ鮑.bk /!7|LmD@PX16H k@eĴkT7A?~WBoO"UɱmK7EIyJhrY`\'G+ޟ]JXIROߧɼJwz,c6iFb*fޘiCL .1>z\ !-X~ʿo^?&]Rݙpz\ gCmr7X.zwI\ش^u~10}u㑛,%dH^zuKI&%IE~;7rݢKK5Phy6s,͘,e QG54>WSvbT 0gw.Xc/u,*YggC`mk%;,{Z8Ȥk0v3ݏUit>>V3ᅳܹGQ6X6_D. |(m,쨠CPBOD㳋o?M \C3FJ0򲴬2ZxʣAi~=ZV'T Ty^Wo~΃jBH)q}~Eo`x 1сSYƫ9Y>M$Ϧ+֗/ wu4ݭӠv$-tm:B(=Lˑ]f/BiHYjn9Zgbl Ɯο9J%k`+p!HA-sPy&d>oݸG$*ƀ<7rq[ي>|PƬB,dv κ0 m']d&D9`a2Xΐ"!HF54v8 `(k,ZE>^p;  x~X"ϓ7>\KR,uH z/h鐆2e`߯_]Po~͇]ԸZ,v'Frq{cO?tamJ|P5unwrNԹrqt:6jh .X28TC%6:MѴ(uA&a#Y2 jhoϼ(K')߹!aѳ_{)<=<֟^3khv2`f/2-ˬ0-{|kܞzR+S{gW͹5*t.*Y1 Z-m2.J ?~md$,ְLI:<ʘTo%F-1ixc΄6 P\P#ckXl {FqE`82`9Q*@0ʠ1dh1$ۂGnX KT~E&Jh$+^n>Il:k4-.0 Ìez۟#%e./2t3_eKؓ# סb eEj(%#a1]F5 ', ᢮ilֆ?n L jE5\CB"B{۳vЎU,L,>f'b d@!Y h{D #qN'Qw)B䅅؊yݴۈJWdl+*:`˂/$2ɜ[qEE޷@T3M2I˄-gh5}䘓Xp(E %d, x0N$\[ I)bQNF18 U}8]fGh/22r[3{}ۤn&kR]7]aoJ 0jh+D$:7<_NP CDɜ5^Cc2,uOVVEolkEr|ɭ bE]r^zwpذO|K AS(!;o| Co< bShKJءmbh36Q}J g6(ݕm1R)pYW^k WPM໲ 0* x!UZH@Z J^ 078VZiv:bi~3[05`<ܥ4Ԇ1KO1^y_?KWx{kuoZ'p$Ib "i,Jyjl~_'{CR05.|!]Manxʝ6 lR G$4IN3^8RB86:Ӳ+8 / X =[k\ɔ88EXii^ŪeϪ.Еo6k /TI_ݩk'%@#P&iu|NqyfEưU,/2\B%(:MeĈR% [(Ҕ8Թ+x C{=F?@B~+ ]3[y2V!N#4 GS\)WL `aΨ`9fDkSf Ɣ \چJ;ݷ+߇ƨElNBs"I -rGz4- XȜI1M.3K B1 c2Gr[bT+2uw<œ}Ha^>FHSWs؆Ea%`\ZɘW6 3ne)mH8}okF6E~̲Y2)we 9!"-eZkfbmJRd2P9h7k a-1(i& ԤkL:qEclqLb7'ۥ4V^,df%]eHk"l =v$p1aq[]JIJEfwvT.Upe 5ƂnNJk sL r2n#7BuhjlR3LXtsoiC8RCƩ~2Zca?6f[JVɹ@Z=9& Oc "SjdLS]#2f-T{0ooz39 o3͚~G M:>;OtYa Boު˵ߙU <~[HV 6zs]{W7itz59:h|IfS@=鉛8LE:^tX]5Paq{ *a>*wO gSucfMgpw/..W;* ^\uVZ{Iu{.➠?Q'э5j/qSSpH#JrcpgxJ9h?,ϗw,Mҹ`yRK:-EJaiAX*XPsQq\%5i]cf4ev9`) eA'&/_~_<QE%$azV9݂u9jJp+8P.Y܂"xVVjEJ#4J&`GUCLWgZ&8јΗ{DƔѰ!Bká>C,Hc Z|HM67oMd'q`?s\/0_ gǨY\e 2'Kkk);3@,#[Rݺ~ia8{q jsL,"\z/y)eFEv-pbQK) hk"$6Dl*)Pcv>IϷPKH&XM{>ݮz6Sd~!=`2KU+p3g#}gӽ[FakCKFбTv WZg[?VRysF9x XM3@d%oZ1'Bq^'L5ax~֯÷~w;/C4hzy@2hN[yʨY[ll?X>vm}BV3}hnºߝCַ&նu>Y[[[#% _\F—f+"*T|/QGiڹ춯k~xM$)}zJa)^;z?(1mNw&E~O:P'TIg; (玎ǭ⠭ɮwj%V# w. {z,XM ? }#4@RR1m>Z|H2͠%MgVH@{\\fLdK똔pŀƑ&N+I, dmd׏dAvy䣧19Gw-{mRbqpĸBŐ1L #RmeA>|k r?h}O =&+t~pO<;/9'tƥS{ަ<@ZOz-Ɠ3^wR{!c^||O/Ci{WR2}n_3340 ZjiںYϚמW(p>J vbnEOa%ox4!"$= =YkU@Iwk"w d!Fb ]we'ߌW^sh>"_um)w@u"yN +γ43-߾q0T|ֳdfz77h_!YϲЪ24FY0ilM4g%·BeH.+lc+_Φa,{@ko_d+z8cz݆|a6}y[I-Z`ĖWޡπڔ]xhkD Z2>nƇM͑8,Qr LXRYB o%Xbqa,.K|[0IR}46 |)lH}w%y`;{{7y/bH"5i;#\LL^`Wȟ=12e\R 8y+l <[eQ4+x=**UKPbYeL5JYi([p=#bL}I0{laIY}KyП #'ڝXcD. j`z3j,&9#M̂VSj'\Uh%HPdC*X}%|dDY@Iv'm-R,6FX"!SiƜJ!/-!YZ]5BR1 2/Og>B%d;7ČiOzockud@'q- 6؞R} .,,?}M_3 )MO>Q-ƀ[೭%L.lQ)8Ui\*A> [̑׏"uʨdcT.Le( Ԩ&lTߛH} F͠UDB:?̈&}s]eV鋦⃫=Z۶op[y^[2{bp/IϦj%^7w پ$WߖbV,&6O>m¢F˅7oz;"'vdrť^&:gqښy[å1"券Q ʷXڕR+ZGY20*ZCoICJbT D,X(j!(uuȆQ$Y&WTPǔf,C gUJm>i~j#]&fUywy<'9_-:Q},^ԫQ=0 xE ȟiҫoz`XV"h嚨Ɣ Sڲ"]qGd<7^+Q3!ۣ ОmA<h}gekVܷA~c:f{|hϱv!0 m]j|>,\0B8K{鴤sq+h[;$q}v[[VڲՖ֖mj/,W'eȟa0bFPz,n߀&դ#k,URVţǢjW19B rΊgvX3k8oJ[/ppP]0fZx?W"͋Yyi< "OiiwNm5<|;ǗFON7p {rĥ/n]~]m ۩ef~hnVLv|ě~=M>wPXs_/ xu6kȟfsc6<9ӓL*330݆ p5%FL#2BRf m}0+nhGMs{evc(IK0xt|E6h`NWrȐR}|nߺDť#b f!5tSdQ÷PJ}fzа0g:ʱBͣKvCp4ߖ?ꈐc5<2d՘Y*l@ ccQL0Ǝk4AQ6b+[FN qxIlOa#VE%؅E+М j0$ȍ!'Rd` >.W:HJ~5A o??D(,FcNfv\7l2 -;Zm4@4xԞ:P%ERi Ml঺$4_;=/͏n"컊 G%m״2‘GnR5tp=dnűZ 8D]0v&l^>!sT[ϟ*{.fB_#r1[;FA6@b8CH[p.ٷkqiNn(i4pPn౫P`f%h#瑛"$.u@7TS^9L'+ٵ]VI6A{`9?|>=>O SnAӯCwtX :\%zŀCtUPϾ،40| S1@Me4w6sj1 Vka- ̩ ZP,}JՏ/v5.uV9Ӎr);/=ȍf|rw}uv2c yV.M]&b+cvFdkAİɷhab#9No)ЋJӞ * IjYZdje kR*a"$ќ)E]RbP,'ً}MWkm#D!~'> گL}bd҅@Eށ/ di 5=ʭ\˚e.ͺr2BXċZ ,"zͺ3#+4 %twbr.4B_?c×Y6KgeG-#L)t<Vj"gsdN*0&1S.O_ݷ5;i YjDPw\m򭷘F.ұ>H:5m#p5&k ӟŻ>Rq: A'2H޲Q͢tH9)`G@HR6 f>Ή @k efG~/תi 7iL'36rp{>_rd65*#F^ &6t%lAaE/jfЄ۴D"\Q6`)tcH3-ۯuB5yB-Zk5U6"1f)׈Gowhi,?w.KcnFugb4 uaFR`PԃZsnu7 {g4皒x?dӉh8ʃqb(Jnؼϓ|0:ܣrI X}BTb5"SGl$`J wy|.?&}ԾtʞR+1.l7-zplM!s]e>4!QFw{IUJ-?=bbCU1nIKk :;g] 9b%#G zu-ȥgj'-߼ fwRWy"19hcnMc ifi8tBǖAf h#hz6J)th lk")Rdo;5SL@ƙEIC(p|wql4wt f4G+gI|wK9Z'$ѓoJ|D|fn~|Աgњ_zKr 2+9ˈPNZA#Dž/26I-4ӳURMUZ$ɺ0Z:d!!#yC-ٖ)DTWYEȫOx3f~o?^?vbfQ%KrlT⨦1WY ɳj&fEm}{ŗJBrjyj_?'I0<汱˶1WhZ ml`6`0ɾSmG0l3u=42]e)=:Hy1; ebjFR&ܙLB!d1GF5%խb۟:aU8ңTGom?Sr2LE}iA TvEVxVd*8ffnщq(/;3˳{k}/۳+mfv֭TuRqscIr!Qo{+l^24 =gy:ˇ_U@GiQ(u{'GٻL}u8JĞBne=ܦڷ@a~ TLA̛65rJF(V!7C67:?Hz>nok*;Ր1ҁ$"ON"ȅہ_a!>zHdl`]v9tt5c5#'2uJ 4;+599ˏ,Z|qSє#:h_Y*tXݤTD#-O*+hɐs76>ʾx1w\04'3cPGU@`1B1_VQlS*Q=l姾Nz4f[I"MbC[E~~xu^?|v?ۻv8kgҚ-0iPrRvyF%m Y:=`X{0H[ 86 %j1KY) [.ߋeE6gQo1so` @gF1^5#x_;d ~}>Ee jkS10e+D}\^F0~R[{R }A5AXl7upO K:gZ8r0QR"_b1w13xOIEʤd 4_jIIdI5ɪHR$a:hAǵQZ+*ئ1ȡ _FXG^bx+zUÇ7ge5vYB9,L[e5Sb2jčiCʶ[wCJ 'YH=kXДዼتFsŒG\VOz-BQ*Ċmh)ajb-X%RUԆ j {'my}<ϭ,mԇGnst[1/̐ǥ\j̨aGo91}EղXWAK"f?}^lJ2D+陝E +ZZQ|Qe9~ .b g(9aP}4Є{8#D)wX8яӚ =Q<&|PlzYMF?Ր8卜M6䀽vU@cS$D[ ֻ}푣EԄ"ތAvx٩8@BѮ*h+Vq{&nY&W W֯f-xy_+P4!Q&v5ɊӽA;f-"Q%U[Λ 8Bb%,5?1_ߑH7j䑍wr!OD a cD]92y4#RMWgdN\k &_X;U@ԡ;┤Wӳw:%?Rdkt<"h1]x PP`BAI`ےYtcNkL?@1W7c!c Nmݺ[ʦ9A?t/@w-O˨sSE oŚḍ`5l@Ϯ".yآ[^] y:<^f"šva8DLalgۧXz9Ԕ1p~㲬]bkuC+1E?PkBOT(.խ os+[dz*E) I\k=bQRh2䜃3gBIذ'(lbV8DW#[Ce_ AS-Q) < Q`&;;=,mrv-2+^5yL h[RE"3ƁߜJĒ񷵃[C= z,+ZH21A^8L6W=lBLjfUK!ui%+T -yB}Gmm/=N-AR{*\ ο=í҄eCYp`+! * y^S  <;7 ޵78k7;C|0A2vNƸr}Nu .>h7Wp]u9ȗBt~i3]5q|Au)$ ϺC<y6vF\gGD֘no\xgt .8| V}sJA # eyK t;TsBW.>fX51o6Cau.Z"YKZ] sZ#χ `{p!^4`=U|beEG;OrFݧhת.WϪ3ucA^.lt.6ExێɄopIQ;yEީhaͳ )8:}_-]>9~˷}rs:3=O)m38qM|kpw])>_G¨VG5 ߬<M$zkƝ#7@̩WٍPKnn?y/]r>F3F  Ƒ)5E?Eo|#7]|#Д#p=D/q n0XoGq+ql#7!OplF.EfvK7@IeӣcR6.  oOb 3R|S;5#sz-/ u < L&N-Wzfآ8 aK9>:o.ę}.警gtgEucGc*/)np~ jnQm:ovTT#`Fb6XTy>ZNv](.|br+w"|ᣲx棷u3p}BRBa3}n|T6rwG#GQ ω!0ZltDxLHgBYў iIH#6ǷBHE? *FVž LHGGa4)LwMƘD5MOYi F4zcjPѰSKu?3x` [*r0uWXQ%u dF?#~#3!& .zQ?'B)@c:i`3!}~tMMH>*!M4^PT:+]8qGʠ{2WWER/z[[&h-.R%/Ubrϟ~~x>[hWpyt|__ケڙ]/ᗺάJTHHƒw1Ϸ9fֿAά{g`7ϊG84C3YP}Ά?>&oj>/y]f?Og!cݪ017ͦmOڌx*Fq7 ZQ턢$P J$vV, %jm.\R8L!u$A.Dzvz,|J; }ۙ|say+/v//!u%b7[P3r0WO1MCKNP.-W]&ZrbZDt*SuMVY(A$ݢ}bHS%>`eo!87(BqHn>Z#zε 66nVNԑ!m'L6~ ` i.JR@BLF3P$&F;wm5qaSLc.(˖V;Z㴔TX>)B I>}n۟mudȖ%: FoCȲcsdك~E མ*PL[k~(8!ۄVgȚMEBYT͇,|ei%I5zƈ<ËbT8F/d9s:"΋:^ܣ06*=Zj8.O<5Ґ'RjX$h纱Zg\<]D#s q$ .>zt)OFٶ~8}_qв3HD̐LM`hXă d{M;jkDRVh0Z-(e@/5D@.!eë&ܭW#Hc axd R2 ڊRc\MzU 9k$$cU|{E$.o/ڤb)q+-x*c ]pǟ߼; YlW@3 f$6  j1ۭ@ .h-XQ<Y7͒'#Tr_Ĭ ,.ZPL1q;ڦR#m>WKͽ!/N_m/k2d^]v ѧhuᤕO_(`8ɗxҌ,jR%R#RuX^``$dWUW=7xK#_ӕ^ ~L%2n"n0lawEEJ)/tF˒Lj&quF/fR`.~^7Rb̧rsBր%;Lf}#D%Y~20DBC5BjOQZmDmfnxyY+g񶠝bo9̒r2KYN8i]m1>V--F)@>#m*KrB`~w P$PRZn X枨u%^CgNH@ܽ*dk5I"Hn>bgzcuuwdDmuVr)>c5M~NO#6U߽FSAlq GnKOmMk#^A%ߍ#%*#du$G$uS0r!}XJs=+@M-C~3${Ľ5XsNރXkHÈw}0RǖBЏv_v~Yd΂[e0-s؏>;avhrYF ήoȘuwv6QXvNv#:av٢sm!6DCCݲ# %o9A7ȴk&Mvm"9gBZ1(LuϕMTěvGb;+jNӇB1$\(WM;o7i}M;bF~N< *SM;oڽhoiG+j*`ROLjN14iMzV{4M;b}N.|NK(w麎8MQ q ȸXCKdXL]`XFUx@ R@2atиG9Ƞ bFf,HWJnT 4uF26Lbc*"CWH,D51Q7E3o֛HjkSuʈ.w2I+E<m$q M࿿}ooOsŗdj> Ejz]"Ngs?wiL`q*ՠ͓&P t):>Ԯc4>۬=slT8 K:7ӷ @d`IB Z!jD6< "]r<6V 6NR!>=ٹczӨQJ53}0#2Obf;m3Nr׽KN'.zK+(MMYV2j|Y * ĸ'*>\J#2]Z)|2>]ʵ{wj^|%w7]1>udJݓݔ<ov7>Ҳ ?ǜÛxđ5.}bOk7=9yE'xˎdD&U* 󆭳>{n ^P( f7Zپ2u=ZX%pDz N2.C-PΈ ~7e4j_~͂.2JNH/ۭi^58 =UICۗ {UU82 c1a853cRi8*&Lq b(Tđ0ʁ\å;ߟ>5,T0㸢 @ǂ[\H5ϘouJu_8%Oէ((:`s?"ΒdRF6 ׁ0J U% &hMf&"eZBh%ihB ,ZC$SR |g*Ҟv%5"7CKQ0c-'vFX\=}t5o,O}9(,rSV牗H -(\5,ysӆ2RK])m9Xol g=8ϓIѰ~ 0kqLsH;qfb62V&`` t8W}v<' PamG+y!vqaNaӧvX"~lʱM!v(D"D$5D<Xߴ'Pc$,QD 2t+84vm=R^ UsGXDQW5bNSbcOԱ;qpb'*aSAъ5ռTnY<dC'/dL a&>Z ˆ<Y%1y8uw_ӻ;?O7O*K-JBd(m$D20Ra"$\մ(-]J[2ޥw ׎oG>C8nHi /lz2_$,ؒñTu[Q_4z}2n&(h߅WM1X̌jےEjpē!wL_azc}ՉRc|i558jN_D۪&3"O3zAnAvlCIҙ#bG!2:t"ce 0 /PO{oU{G2\o"Ո9JvOE̾p6}%o6￧{Kpe<5>ښg|'ryo9wkV?/9 ?5\w魵goǿ}+ww@O.r3kWX/sNr4o4li#)>57 SӇnZQ7𙚭ڐ,h)ᓴ1>9czRyKG }נ%LwJWlLsH T;GuT&Q%U U?_5xYƖAic~ sk)I7$I1!F LG'sI:9t.bw\>Qt50*7Ϋ6 ˾]~B QⲵX2U!1Aa/=76s@Zb)mUU,ë3fvNzk6(nqD, @D Xb ~15+t-}\Uƨ8&fI ֊LLg!D,iS)Xe1G22M1Z;DԄ8W4`IGiA-ฟtg~=LmOb]hCĬabzTΤ cS;(.Qzam#-LI1 ԡcƐ"5K2Z:& :. r"b40qD篿G6Y.,g呅Sjx[N0v?>xdD-A/&ܤ:SJPb*Cި8\ܑb(DXIϷO8}ۿ{_ +s3pڒtCȍNjy8yrqzA|~o^1?i}1I+go>CRKR{fw%7U ;5"sLZ͠gl]uzc&* #!{.Sް]{ sSף8М>cz\.~+ؘ4Ei~Y5!2kAk]\w.ֿW `d} Ѕ'\m'EJv)2G,3Ly^Γet<<.lϿ)/'t 1=?M~П U QPPD&<2D5A<=b9(GʲX' 0s(R !pƌ-`_Ә6@KxwlDY>H+Bg CW0t CW0,078+ gpv gpvX.k,L1'֎ (g i1d*:̋q-1J4=?|F vsL%AMFP.y!B+|~0`hb}5v{2:vۈۣ.%`}&^! EIK*dN݈V~nBYÄVǔnvVx< =bncp+00\˭t큖Z#@K;;ѕ^Ibz+s|S-Y#7 d+[7E+zS(P:Wʖt2,V-VBLB%I+[ȪVhƑmC u<v|V[ѺFQKfQQ-㠤ζwk*q:NIqN(<Vp޳qe ~ow;}?,YUJii+}Ϙ8`3f8c3ܹqNxhMQ#:%f=r ~lwϐYڞNN]e.c\L3^sӅqc1!3πg;ÙaX"9˭Iqu_wm^M: tį|osfiq;QzҼn~֊QI$|?Skgs3ssf}kr:Da|z$:Ugu餘a%$d3wh ܺ+9Wr[uS aTqWRQ뮴 & jݕZ\<&wk<sWXEǎk hMvKTښv9V@Xi![baEИc26oښvim$Cɴ1Xe2RiiMִ[-pZ@L; (~[.NJI+b\ ikڵr&yk=t# c2D#&< l-a i5')rN;2lk>]i*at?ogIF"F"L\H c8lRqdx5|M_3`zt P~.,c) >O ) IY .GRPޛ]7c8%>BGhk .5on jh h02~IYQrY+~(+-)OOFs`tiCC?E1FGK1?QH4\9SwY$kI%C3}{yLR_63+Ov`nX!c,"3'=f83{FrX4~Xλq9レ_mw#P[&vErI;oլKRVݭv+4sـRp>O I5nbOxfd~fjϑREkm(b\j" kG$6F9Tm4`P\南dtHlc a, X9UbC8XpUr2ɻ o?93$ 8t"43Rj lxv\R+YFOEz!Kkn2`uS)Vk> զwfJ[W-.?$*dn]$%=7Y&(Rc[|F@>BY#>BY#-o%)DRN#u8D 9"xZ# e5nC^AP8! %}'U:Vl6]jګ#w XhcR0˱T*JAbQ!tsh<+Vɨ]},R:>   UBRI0$(5pT_݋:rkD`÷ v#bzL#A)sBT9(7踥A{C@25/wmc3Ud8n}< lj`3$EsǑMC9q& s_ŷŗ/_k뜬kAOn0C_jbnWE]fTc%&c[%#@\;s[l]"Rޮ(uoj#a.ŗi3H74I-T筩9 NF`z[}}; S=/չ1՟>{l}v75piߺ//~? \~zláܛS,/l~5{s`x|'w Ϟ^|wȳtOOҍybiNAt漾 l;§8 tWL?6U3!UQ`6do9Ee]SHI .{ޝE5dTIqU@S cME3v]%T5E5.q| e-6[.ܿˍv{$(GSe;/LZ-Ϣ~}:[V =>&y?\4F{du6)AQ5_8k5n}㧍{K4[;qz[}R!vmk76,4+5_l.WKFI SibyBeg5wD+pwX[p䈭j(M::7u5۸g`8E unZ{ Gt+P-l>Ub^tg'⦶B`^`<#p_hOdULH~. Azn@iA9{v b3 ^H,O:2)m<"U ("Zy[CW^vkBmRU2VCՙ*',%P2AMD` s{2'=-a&ފḟhӌ3ƴfRܡ='.u5'\ߙ x-q/ݟ_~9{"bOrNF1#ƽ}v>^ξz*LF|0]$|u盙E3.femnirO3}1˼F%!Sђ ۨ1e@sJ4mU+N3n=;1}=xeEr&^>OfYL Ūv{Rrn:Qk?+,l䩙D 9X+r}`Q9+5Aw?nH_1Cڮl = C/(U㵷wmJ^[YJ%q&DopHb`: E ӘshmQ0͏>*Jq cĵʪ+gdR}v I7Eqt=kIoID#m£=5fbkjl 朾PG#!2=V0Eƃ47"9lOA~|sN͞JJ5߬13LS=I ˻ҏ޹irn1wN *.jѲ 4.ITJ_^R3W_ϯW ${knaΡ""B4Z@?"oۯώY>/Kib213@h\314/Nߙ-1|SiѮKIٓ1C쎬Ko~&5bŐ3^%O*IKQbkc3gq)Jd}Ҧ 5tũJ22*P ȥݡA6M}N NH !pG+`K*f4Jr@?Tƈb/Z84Kư}+ڠgp!4hzNPA~u-Q*ulnkz 3P4F$* mYu/|>dڙ`0t/N^$S':qyla )E[&v7t*,dd4J4nŴr]a7$jæؕ2v< 1 w eI 02aXm^gȤ%#k1M(2aE v_m @&q6t fbqZSW"DZ"i4sry8eD,#i;rnHJ-߷%7U(m|[H۔J0ψO= 0xRCa%HvT  z؃1 ةLXJEE1CBT檑g3jiZ=wkoϛw@h]$iwgŢll~T| ItS$ VIUi>9qfEGTMEEh 4.(gMܬ-4"W"epŔ0u֠I*r -`6*͊ XAͮM0Ħ0xV RجhyDo6(ow4N6>$YY<5#ϋl³d򉫩@䗶姡ɽȵc%M Vg&b)bFdzX9{T㌴j`֜Z{RFjoH]C; +3Ib؝$),ϕ҂siA޹ \Zw eccl ג9,K0ǜKI I e 5bnOcLAt12>(|owđ aPt4L꘿ 5.ߔuw(U _2[{=B +%AuhA J$I'71D![vml"([j3>9R`ǸP~;׀ח N4-N{;b&C,gS(3XK;_YSKh;㥔ĸ4V] /{v3H{%T 4HEM%ZۿP ځ8%nx%Zl`?g0 zȞ罦skKr+&>q^IlT $"J% JʴEڈ% ICeVHi93͂0噷Ep+'GNi83`IÄcd)M63J'FņF٨v~ms 4F:SF`AALXhbh ,c39gTD!iW"#B* IY9~-Jj[&Jp TnJ7ڥt7.Ŀ)^KZwg.uw$bW^1GRmͣ˿4̋ztOf'% n83 9K+a]NKR < ~>L-t窐NgATrN 5^;{ZS jVFï֕A ~9%1[zJ@I2kXeiVsB3XK ͒X$I‚b](Χ #'X l0*WĨxa~ A)]6<~\%큡u"(QqU$Id&ه 4$Jdm320Y"L`5Be"XHcf@a+NU=I&$rUO.Ns0oԥlRFE= .0ۺ 7Y >`1+Zƪep:,VIc,IM S2èJ,$KE )He ThUXM0,1(B:%ȄJ3crSicD5EkyPC,^&M)؆ 5k߹Anꄋl[:T9 '1,V) R`[ ""ٱCPnP mJsg)^ W+HN>yԾŌԝ%KصQUF|:(H^Ol#v""||OpX4XD$1Y{vrM̍~Sb$Lus*ʰ2-@AF*PU :T3xT0HnTrc3҄16qGm&RcI 43 V@ʲ 22lPURXheMLBtLZm~Xkn:l@Zc ,c\Ƀc苟} }P3>`vDE& ~͌m:=; R 6R I1$c+f&څr*EƂ5(=,`ǨbL>,ۺnݶu7ku7m=<7;=DѬwFSmװ4{KF#$O_6ŷ\h~6c/n_?8լ-9Xi;ms Cum{dymW clﶩhs8AL=Qhnp*&t|0yM71׸&9̬Cjai2ׯAU"QIj_t=%z;Au4"g/0H#ؓpD`T CLTrO)eT1|Sh߃S1>ʹM[ܱ+tjDkz_ {yyƈ#xU'ُkziBX|կ2]M'~_!VGPN[R/*!n 6u6/0Ki9$w+ͧҫR@9.j\9W*p^I1=0"OF`}1{2Z6*[--;.^}oó|y1s9P=N^gb%WT7?Uz}>Wv!6@SwINNƙ,(k*rOssR鋏\̋Үl4jo ZҞ(ejA֭Z/6ui#N&s(F-Krot/ᦛ9ַ^pJIuӲC1US{9ZEr:z6ɓswW*6nƧgKFW49wӗ/B/^K'6e>ܵzu1}Z-lf6[(qf%osȂ*~K,lfew3Hߕv')TP$hdQ&,.wkX 1n|; S^,6~Rլa(SU]FUkxwY{}]=v l4~d$ժ%{]%O%g%{]%+钳|~@Tqj"8b$ELD8q&u/MTd.1O!/%HSyV6Dj4jqR&Fjf=B`(^ʇo]+3|{'~ >FCAkLjE~31u6!M leX`mru: 9Uj[2/@WpcǝUl[^=Xu/?]foV #?5y겫u8f.ʧ.0%,OGm W^?n?{|9Lz0%X4hȈfF`HoR3)kw R9>?oohir; yJXb]v4P!C?剹k'CB!ǧ1A*i.L{od og>/zod콑{}~ıFjk1.odMXLzod|y EXfuۖț uIh3o8rNU=NgA34QRJ9>iф.h`*MH}Lr]hn4mv a. 6}RI"X?8nPq[7j#Z[GU>Oom[ma(r#Dˮ[2-ϋu!az.(p4(s*TRWEeW R){2$:M4\[Gw#ITi e i똦&ƙE&G`n>Y"뜍Wĥf;b"/Ԙ_Vje{[ɯ?ͯ8-he.j̄d}0R) fN)0L_`/0S-M(0#CqLQl%XK {q/uhy6! ԏIJ,#91i:̦ϞQ3bӣz u0ۓ@=2 >UBXWu* F)$F3H J]L{70/y,}M 5BF HQ"Ć& K֜U`)Ul4v c^ #${M$O$?Hk&kU4vl4IcU(DG$R(U1BiLP{M2LxgI &,Ii#K •%""RG?})uKŒ%lAmwᷦAqbk2jv+.xwBKj;v(pX --:lw2](RfCպŌFC%J DX `ZA |Y=\0/9E% C/A;ׄۍX`~MΣ!W#t_ݳ*}Ea L|pnў9hvn}YϲYGxG+Z+|JIº~z,x~uK|M(גt`D;_8J1*.kkFƀmIƊ~ 1/'37w*8n ֧t?c4so[6sXL7 C4dF1FyXfq١{P(_m01p^eѣUeV7sۮ$Imx!9eTZ;o֔kCKrE~@dOUpEInm@ңޯl1.SY>Udn 8ӵ=f&Ͷw4U(8q:͙l8Bs>D{ti|vW|"1$RWY6`Re=%тz_?w!Fc)cdEHI66ڴ0JίFD))Lp 7JX$I.Mybg @THfhH)a#j9jj_0Z9IXjQ-:!4LISaj23!X@86XrkĄ# YKcEC& ۍOl5/akj0BYBgƚ*MYD@Z NHBIMJJJs`+GU J$L{) &,581ĘC d3pxd߃֪xv4 4ȂE8?@MYO9"_W;;+$JQDiƑ4)ƠHBi!pHTXj"I0 R&q&N>Nj]f6cC ,̷Jz8_읃Ž+4S9GI헟ꒃJMZ]f7k?d J+q|J YY@}d3KgK-Wbmp>%q 6v4B@xYݾ{6Cr&BbrJ#)KjHBKq $"1DwD4+q]A_ƚjVyճ̴B!)-k'|4"œ|y3Rܙp*5PaF蜋^1  &6R_%b|r_e$+!ǚ A$:?`0rd !蚘Ʉ<6H/[_=8=`lz9ly +H\OO1<8{B{ʟwàO^ _eGB~0Ӽ$_] ٢ƨf S܄m `ݝBY` Mپ87Q2[MG|:ө *>g+~tBʵ#% LEDXƙhCeHdv'%t) 8"O- ւ,8):%`=[.%& ,cL:4dQ֦™ebП5AT:3`P$K |H2%6ھm[|w 6<> ?pN5|Y|Sږ=SPB`[!Lk.||'H2)D.߉!!Q8w$ܹwD&S*|'T^#x#2߭h[LNLwWBu` p'QZJaR@O1rsS+rf7Wp4 ‹!?o YX*xVjh3⌖ׁ`&eȥP(yKk"(S &`*T8V@JnD[ v/  f/~"* UIL΁?Ps%2vjB\DIHx6HJnۃ`-:[k^ &;X$ɚ{^3jFGu%̦~KH`)U<|s,dEUC) K/)MqL"WnfY wj$/0ox{_l:i%lXE%DT; m$I^&zA}8OvDl;;=:enK) $,w-QP_VfV"Lkqrq/N87 ${H !m6L0!dw# #{=R%U1͘*팩'4c];nz,j){"}?7~PK_yƐ?k~/),l~Ќ&i+Ueu4 tk6c/7c/k]?5]6LT0VSg)p%ؑhfjͣwFH36BZcB$ -K` Gw6= HB롁to;pEP=< hIBov{{G2Z] QZhSu<+V]YvSoGۺͷ0[*`T`C f7RtX?u$ YHo13 &"yu !2Ugvo6_!Cn}oAN#!)KT :_tR-Oz%$R^^wI/lO9~ֺxpV<4[{ԿB;B;vЎbykGWe]eBJO8s 9\sV;cGHRc=YɕS1v㗽9ZoޚM*aF?uk-'[etm +mJ,B)w"=nze(,¢` *`dV}Zous9P(Mi5?ԃo?+BP5i/,UK-4LQZ6J ЫCRiRf=_ Vg)B/7}[oވm-zx1eiCO|t~O ˝ޕe hZMһFdHWq$MX+ E!j'Zq\30JT@)H8GR# [dYCjW+>(Z!'wL93.2X̕ `0O1T^pT^ Ea1Gcw&hYcL"c9R $|_}|sy|G(ǛPCp%-< N]yΠ/dMCIBJ==Z FT(D[L ;}) AJhD– # ^&bd܇(𧽍RU)0૎D`k_smF)SAr;(:U{J?*I>XT<%R`%nq ,RZfO{&{a[fw=ɚ˗c&<6kjSsLHmAN5QiQXӆ%*{{As9sY$$x> QD>I0 EC0mX glRo;3m5m&Pc'=5! PЂXNƚ`+à^i'ja#CL(ܤ p`@*u2 )#XKp4=) <0L'M1 U(X R5ڑyp5].;-<:J/&HݑG n-KA9 '6WZ^\R$-9 *H;`'JLu;lL<]牐1D#>/Ơ2`!;1FSv Jfz{J]K==VgtٽX G=S>5H hUymps OuWwu~޹ߌ8I)K \Y+:%vV?~ϣ-'+-*p⇻UX65HdٚҚ.f<~6}vч뻫l/\h1"*yՆEսxxO!VTT%l>lKaCj觋#P>sz11A ;ݾ{]|wlI?-Ľߝ HaRoI!x~勝UbT ۻI%aHlłZdpc:nMYGS[ w1!v^n}&!ѹ @AdARd#$pq:7$y(3R ^ErvG`J4麝3$cpfc^r6PMV-n1=3P k{oզ.t>i862͢?>~/WE_zI)6HWhE"q^``LYU<>еp%U+ڡ̖2fJy0l EǪ/ݿDD@Zi勉3g4bFLi=9Z"F ]rbvCb/%s"CuelmZ?%"g+,@cIMLM\fU"LF@O{ƪ9S+7zxm,D @ZȂ E*H=󵪂4pL{LiA[I4:?GE c'h""=6 4_ V-IW.Y2 ZʹPr-uD'M[KtZeZW!!_fTeJP#{Mc9˃:&Qe OeZW!!_,Sx2,^n\ %.r 5Vv,f,`i7m`4b!US4 P-"QA2B41,y GZ츷X!,6_d-HJDcYM!Vzq>1(0zE(r#t&(\t!NI \ NUx l`|ӎ&x?$m'5U aA V{KŮ1PXVpAi# ^RF < ` Ø,`%IŻA67w!#J8YY'TX:ݎOɕ\8JU ӛzj/7LTͳ/2+H{N+nVօIE\РInE"~K8 l)0A \VZuƳ:Έ$aLܠ錺$@f!Q%8hs\7>[ 6,qEpD@X UL֍E' argv㸱A/AK_yS ,,)-ɞd\"U8N./<<AP` kA%\Y#3kki=՞XXJd>9gQ/XhDp6- AF |ZKsa* +mS:o8\daQT wE/ZL4Y@Kjyju\8䴑D@̃Ǵf8h8%-{$zBn TtZ!X=Lѝ8~#=lAx$!POhXT&84{z#8ND]HKvDkBv;9 82 @mTܯ@cj@C8 "m!30,d` MPaP ,FN#03^ZT!G׳?߯Ӳ:Wbr"v;zɢ(VXVy8kE'U4x+:O"nT՛x Z }iXW&qgO.u%BXSM7:D )>$Qh$\t Fџ|%)I96FICm:{<^DީsD3&h߯L}o5frYor'_ƣE%Ye'#Oך=iS#¤8r'"0:9Uri") e#z~j4(3 32b=fzOCs$ -#"W'T=ɫĽWcP0.m ,<- gC8kiFeW9mi(T2Aj<汕ip٤ϒNm쎖 aT3]D&Vwn fWmP vqS@p5\lC3!ƵօM4XfFȪF`Y4kLƀ;fֆ~!ct7 H;-S:qʵij`eǤ{rKh,xi&PɌ]1l>Zt0(+q~Yw/>cǼBBC-$Ju߼<4j!A|JI!Z@ 6+ #OlčBL).CDt94@lkME4zi2$ii9ÚR &/3(4[lO $XCp v Y(ǹ81L ,3q j,ҭ2aӍ'3 -ܶ-Gm(Vp$ck֮|S^[Aԝ5-NSHb{mp)*govl-1FvΙ 4[6ڭp)=], sqUWFRtQCzF΢y8QJeSc7$Sg{ A>~ǧڳOVp-ָ_-@./$r)uf%./0xW r˵mKKrkp ΁@31mdv,*,xݚz:Nb 7xMnt,mcL"մ7 Zh8E-"e &Ze6cO͍pJޯ4v]'(wo}W;H~{ǻ;G[GM~z{H D]e'|C0 ڷnWL>ŋ՟ެtsno9v-#,_>O GR `WsN;,SGHp/7~(}VrȳY %(UR}WxJC-h,h.@}{Z|5 1jK'q,'XQyBXm1%MT9'p(ʁr0hx%g2Ogsuc*|SNrj)ӣ*'6qئ>24j "}J[BcH{/ _Byf.~$ 5f~PlV=irQW1Vbn^:g~}@K~OŻ=k̚;%q~9"UQor4fa!8>QM~ Cq1+-=)ajiw?ҐPaP)!I݅p13BIzTu'!$f"RID'LIi=3IvMzɵ`u$W{?<@{~Fi%htgdh򱗊N?F5ܥ FE#`T㴦w֙맫Sq6쿿yp6`pSNs)§V-s6ƙC g> auR]I2 eG¸Gu3U"S49dGE%b\ťV {T~2g.>fL] :g:RR]X[0ԇb0gjwi>*AC&Waۃ`HҙSQ a$%ɠټe D06P^ݺ RByu3Sۓ2":X2"\RJMԓ6E=HmLP^l2HKJh URJdbhG,ڰ\n\֥i{L+g ABK^}9ˆfj]̀A3uϺ]E Jkw*Fs6'/|.8-_݆[]Z]p`q ™#ɄCj iٸ|k蓵qΫo|<%gJ^)62w)}IҚԦzMR+trWU%t҂_ا ]td+1 $w-؄zd&{ǟ~aB]`aDE)`UX7jOS&@U!gIN puC~{_7)훵tˠ/d_ Z(IBza]heO"=zYGA # kآ%/z]h+97L}䟖bK 8BĎTQhݞEQoJ3J%PQ;6Eٿ^ JsI~_tKEž5;5vghߝy$lK\=5n0}^uYx-]M/ hN ?5qz͓ьפSr>54#5Yt13F%uHs|?DI)"IEPQZn^rH9OJ2I&WwQaPiYzbG5E4*L׳%#59>DojʉÅ/ ;bXkg}` E3sc>}W _͍;̻E :Xihv tx''AO)>Jnyvh1H o i4%-P`3, jZ(U;&Ҫh>}h&jABo6mcӞo=kо.]In w'8 ~I+N~Ec-Fs%`)'T9p ּIJݹ eqw⃹[jӱ95~]mK#G hя#@G;#N־JNEjܛQW˫aAs(ܜ=Z2㯬28Sofs')ꋙ+ q6}fD! &KOqX݉vdM9ҽR:itǯ5wNjtSDX}r`?Uc.dpnkW8xΎ2M4:O%Z@P)i@ĐG7>څD]6j=ZvMbDMewkG6}/JՈr[wa=}~5ǰJZt2$;Վ4Ő}z/`T?Ҕk>_iΎR;smL($k$T3r|>DÒàhn fh%<"TtT[hQv@Qk8(pFu2`ںܢĄnKQ1# ]S*FYŅpf喐&ժK&G4Q>BjB[0#N0"/ЋK;^bLm&O-@0q麭rr}0j =>D+N8LèmZ'm!V4̋pUWֿ˟T&Sj+`{6ߕ]o/׫X%TReG iRkbٞZH$I6#tLJ ^>]zQz+#_~펥'ܤ (fIw.{2 }҇giH< O=.[K/?uHUB>D @aa+ C2_Wn0]`MeQ>"qJ2]ϗڟ˨̸ Eh+hlu(bRmf+Lɵ2 87ط.o>*,ÇX1=$@(?M)IAe JDttӥ1;uAV^jGB oҮe'ؿ=>:>%(|'Wm Ap 3+DaJ$ `0\}bG+zѠl Mߛ/KS\M ".A`cw{ L4*EbMTXf_jNu NȨez#uUb1qxgq &42/pStMX ;FhjO|~x+NA,<.Rmx-\>r'DF,=ekæQ5TMS&ಥ9/,9bY&Ш(' 7 p'>/8hvZQɃ\g" ?Zq(]\ ̃>g'FuA_ʱ9|fxцV}rO<7mK¹Wpर)_A{<EY۵mV [,j4}/2_*>=Rj[e"sg4SN(R~[Onwy;X u[#1j P|2MN$VnGX=zgt?s@΅QnW5T־u; abrCoq* A..dZ8]ŧFL@\ F: r,J,?ҳwcA(\V Eg2 &SwUBƀ0;<$Bv*1 uZdL."8؅Te6̺hJ0:tcG,E^̢vԴ좲*FaɪބwA Ж4cy jWhdo໅FlJm[kDsa &d8t VKB'B`(xH>~J.~*=]FQX/zs>*83JBx8ppGI@ \ByP}/ TAC (#DerR3jL4HvKSDJx56z;:n]jǺi_KriaQGX ̢"kt}{/"jRUa8dzRސ#PY?\p8_,Rv)|SB;DwFbYwk_M&y=i)+Umk|Ă8Y.k.6)( c/K(,>oPđisۿ>|Tk ɝl˵ĥ a~zP%T7YM9!>C[)mC<^ez y">w9 pǔ%5ף.To/&"Hh+$):0IlƨY4%J5bU#@2hs6ڎ-8r ${ c9"iswKB CH $OC}uQ}eх(pʫ4 xeιRz+^Â*nHS5 Ks<嘉D'5,+*L8 Es1_Y^J8b=*T 2u4T-WV6nm5kO,}o_vb&6, gbs t f?jyH(NӑJq5g3ڜ1_=o:o݌VI{~J=Y-"ۆV\(THf(N;ՎN/a]1`"㳪e(xO׎b0^sÕJUXDb @4f )8|sϲ60=^0RV,OIf3qHf9m39o#w9m dnWŬ5^l|bJ,{y;?Nԏ:د@_0q꫻^9q3R Ռ2甅9u9 bDst\oڲ7m>i;A.!ekWc~- m)Ft A9H3;kG1} 8+ɯ{˯\xM@dZ>9cf)OVQH܄A_e&=RFE> 2BBKJ1v1P8%.R.W O'rMB>Oͨ4-3js2 29C+#bH\sI"ު"ZGeSD];}9@@ d Ǟ|r0 ƈJ_(\yqWa0BV_A b ai!/1j}"vwZ]1~ՁQ!h;FӇZ7Χck}MZ}BG|YQ./mY=̅gmTSj R\~+̓`ޯW /E_#&h"Jme{Ԛ”^ ߌ׳*vlFLe .Wj@jTh7RbEI2e|+" ELQ;r#h7WWկs]Jea=}}@L$fu/> Vzo ov.["Sz`g[7osdli>%q2ZHU]/'%i(R:S$EuFH(f&Wz \l{&9 ~,P(g)HUvO%XBB{mk2Džv >:'cS G)jkՋSs@1qѶU ^nfǧɊ0qݾ) |`m9۞0 >Ag`htS*v#)>$ _gB/%#i}6@1!ADlonm*/Gȼv$]*tzR4B|\/Uh&H=b lͤTӧUX} )?jdrޠfe9_$*j![gIOH𭁌zݖ!fCuzd= !$!:˿rɔEr`:T x!DM`wVJDjg<8yH.=DOѿ gy:XEieZ1,^5HTNg:AQl!"L6 QM"ZujA6Lwq+1UϢnjjj n(G.dH09H!')!Һg6Z3 ͍׺SMV1zڸx(8>>K{-ZҼn6&C1[G21W!@2|iQ445& 9Ibh`S@JiQaZI_K4Kw^XZHP._:u+! ^CKEC3Hgʚ6_A:ɮ`L,{3c+|zPUfF,MMn`UYyԑ(Bd-C.l̾k ghe#ׯUlloޛnЏTT 12B}*-4'RsX…ܖp4U\ iwyJU c0OX"+ dRH#Xx "SfZ\W3ŵ WW`cM Y.zs5 z TɲT=H cĜ$BQ8!IҎ%&25Z|&1s$7D+M$FPLuKR8؀z)o*ڏԄP"O4մoGnխh&O{j&&~f$k>fzj7AS.GE"K {3 Oq'!307E80Mc rծ LbAnYIEAH*QD $ Q,4EyLuxib>HMU̥c s'%i kMXPr(~,T2IU'%f‡TmLM72&Ugz"'iWK)GLSb&SL5ի*JG6`i),[gh؟NHs#WtC/HG =-[NǢC^6,IMեwMC";覿bD* 3ꐳHsJM_/7RG#5ѻ.Ip0wm Jc! & )A׌7y7B{%ݱn<C[lZ &ڊ8.QxBX$xL8ƪM%y>ogdoA_QVi~ PVɬjq>V:5#G$&diR싗}N;L۵QP! m,c+ a JSdQ")7 H|-f^IHAf;"3yg=P,)9M,Hfe"f%J3nK&N2A@'Av= dw0ۢD&Y^0^qaZПt2'Y 3kADW* q-cf^÷YmG>E'|lYُ~8~Gm啌s%zq'5%X!8<}94602JIG6oj5Ω՟9)oԬ%IÓL1Ìac5$u! r)X΃S.΃SΑm|rQ+BJFޣf`Yi稛 TuZk!ǵdIDwi>V#kAwAJ˼^ftX %Bk( ʰo!{lӱ Ҡ眏Gol@O *5 E{ww_!P1MXиI?B੢l!/S?C0#i$`~gts&4&s.=ɒ`e4#!/\D/)8W_cU[S N>hS*ZU!!/\D cuF{ڭ)UDiu Li)[cBj6$䅋h#*;׵5iI'P5Š4v;zvڭ1U[EL1~s])rڭ!Da֭xicB+k$䅋h#Bִ,vbPEtQGp xz]ƄVnmH evˈ]O$O&H!$>! @y%ދV_ Zt6F]{82[ewj̇Ս~|WtZF*Ѣ D)+]L}OYU(uavʮ,KI&^M'.{wpMF=F6ţRX\R` HiDuUCJ*qEjːRbLTh_*4-࿇Y kt 8ުE2C-8݃A^쭶r&BM2c;O%EW4 4Y34䭿t4&4eE4e PKZZZZ 9fBs.6wŚiP R8s0S—%F=G)au~9WJ @o~}Uҥٜ-#Yyk=-\Ol7tT^ªʐAׁNs'_i׭Wc89AYI>$oӛ`oV*lpBJ|Fduኩ1"__kM밀ra($9t-T=eH>I8pP6q :1O819,1_D'8j3"QSRY6Ӗ r|;rAK*/H}eJR9X^(s>Y;p/\iw2[?g'At13xkYlxL+)_iԱ왴N'8NbaTl1KtKAFE &pL L~ӎzQq!r # w~8A5|oɦ q 빲vBM[M+!oAJH:"Z1g4lXi(Q*K %ly$FsC2ZlOqW|^:f{ ҳe1DGMSUF.©C+_mϺ,`]~!(ߤ˰ҭYd\Ie96 #(%`ː $\hnN"f,N,MZZ)e_Vify A^s8N8#S($n)M$B q1fGZJ/ނQ/QFx#lB%V[BDS$AT J9f0$a,HbΒDx  c&9']6| S&*8 bBΓ!as<ԁ"n 35OXobM Bc*4yu.!gw#\:DL_l%u /ahXH@qcCf.Cbj.{:0 c!8|WAjj!hٻdWy882{cINٚ4"5/o7IDIKsM``h*S3)J) )g\ESc"850K'~o:;g {6nYa&%zi ur8}1:SMߺ#ی)2&[_sǔkj8F{jV<`HjLgFE:7a7>L0TTr@!@QztKᭀs]Uu0?l%ܿ3?7WoTeoc5rU1jFk\<[TO"W[_9B?# ݬp1SHPlW~wL5U0ϛ7׿4|y[TۇX"h`rV&p[㭕*ejs(IU^PUp,}K*owl.o~ݏSP\ȕBƬ,6#m6]uao&Jh G52W\,e^Q`"]0-};c$Zr>amHOzc7V]sf-_̒/382q w-\]q׀ܵ[UQeOjFM뫫C*/ <ץ*[/+[^ysIX>7Ol k7gvv%9_>iF}qUX[?9TQ`tt AwEd94߬ 9rrTڑ"@ Q*6x [Oy '~db}*nEKSdN.|]H_$q |B|'Py`sQ0gLtq@4ߞJM[?QA/֜'zIOrCg,n^CC@'w(FZ=eߺjE(6DIB Yf xgǺ8Ca,+P373Ԇ"Ǣ"t/c)dzzj+5U)cʼ[NJ>Zg͏(ky5:PW_;l F*f0՝ʖ(-\e(HcXt2e=aoZNN"8=f S/\wjn 2wP4<  ks)n<-[Lρ攊~pgjFܮ`g HDt>Jx?fg!xo=d!dl Ih-fgσ Rk5'#uYɖ)U>nAAwfDz!9=-W-L68GY%?b5>.4 6_y!f@Q' [p ,zOI8xU(\ëhio7M(SHح\D1se6d8H19_vG.:rGvVr;pI)ǹH2tdֽK7n ,99lWso,_Ip[PggZ }ݹYㅈ1 i>(%ekS=8[܅?>![gXq ;%q Ǝ=̴Ny= N؇ C >L "lP)b m rk@*Yp;:*~Qec{:e=k+T@w;i?CȿWy-- <,Ib"]jww-'[~RU m@-K9}rC#G8eR Earʆ `TΓQDPbYtt2ܰB̡h@>5I|4YoD`7[ 2L2wz*NB9BAdz98S]B8jBĈg|KRNB H&#}t?|AT = E~W\V$[g'MjWk'2tx\cǛ RnIf{콫4I̙zT/C;:'[ \@gAmSĻWmV7"xBIx(?LNtϿLP C%msN[ 9ʟ?Z[BM]=l߄1a잦̛SF/R?^؅"֢Xn%"iAR@ɋ3%)`q|{Wv3 MSaZ| e"XDU`xyo('$ZZM#[+eWF\elM%`|:>% bg~=I-( e/_7awCGӜ=[у`D`-r̆s: k܇u0*Y`a7R{WN-3uv`#2 g13clU1{"\SqN 8 s昛Y9]ё2VUy caas#mz}xO* ݖ|[N\ʛL7o^6yjFķb z*e-Ps$x|#Sd]@4A?M/yI^xNuk>k^!cP@}-zfE4kIN9vZتy߀3Y[]që }zclsk;r>se6`C 9oO4 G? nfީ7 sJۃTg/D {$Bh^ }MX dewO(ciArF<oU NzdX^I1lΈGSZ&I8fz'®հCZP%9Ïذt` puKNk3!A<S'^WƓ%C wRΣW]h p8Gd.ʽ #@1#.Wxplkq4N:mY  c^7[2-n>?~2uA"Q`-W Zgnؓpj#o,PR"OCL+t՝8 fFMH3Wxjݴ#[!!ױj+Ycy ~8}J;R5ҝ;O}:%1b 1炿lVIWհFpJ㐰, %]D jA~lv_}Cp,Fl["*#&轚Hˀ~A aPzj 88M?/3gOV^j)=(9Wo%XZ""T:oQ+xzry|=*$ԿMjU-@є|S R_!:JJQ8z a. GTu]ӶAJ1bd@Ë$BpD$` T` Ƚ;G{wN.0>}sbP0jQj mCv'w=;>8_1 `B$a62ʘ aC 0#r IGK~_CB*sp\bQ,M<<4(Ѷz64p\jΒJo~jh ]jUp=GL({۾gQ^b-٫Ռ6˖ڐ_\[VI,'O\ĉ_3NUxQ{7]ϯ r@5pmŽ3tow~^ji3ܑ!;55 )  U=!<# k|s Ja13cL\ ˹@靈|n@9uxۺt-XPChT;iP@(p͋gh5 >Zmm ꠴.Vt`S$.[aWlօ=Og" _Di9RaDiP: cơm @HS,ꅲjM8PMSm.N؁oGj3wot0;DBu*Hwf_(6$amd/I x<'!., "_zpX`8蘰uD"S N{WT0(݄MYn8M b:s_dQg;+@DHWZul3HUl5{"hC)hJf7ǚAK8f#ۏզt9ϞS$1 rJj)h\DA~ZdgA dǪv*CkR _jr%Pl:_ݾS?nC@,ZzU`,Jf+Js U䄴 %b^6 ڷhH\.' &h\cGl+;AyZ7g\ͫS7f*4~G@M痀09Cy0dPx4d8פzNDͿm"^&"'۠9ћ\J䬳(n1I$.ro*~21u^N)f٢D=N:݉@#-V`%u VlH[f?wp, I6-p3,9wzҠ#;`vkADZ~s- ݍjiSv' U4{~U2Ct"^U!uJKS勌ŧDkeuXQeÜZj3+5Sy8Gw?H_1!Ap!Mr%Et{=d}Q${ ۶,~U,VU6raUsˍ Y,5Mzn9ӈ.xtJ$C[ ḷyֽaĻ+7/W4 6a'O?08f%/**.rұM 5 j&Di}3"ce6~S4؈51%U)f΄%&| ǭ7~>/,뎵ump\qif-&8bo~tn.R\Ǻu SƱM4XkZGNOј7Y_ 9BIaZsLF\G~8};R%QR*>e r/NZ\NGժ)_}q/\|{nӯ1UY8j(\2"=;JJ iRs-a4N8i_qUIԔ<[j\L)=ʠS4d\S9~pBxML=ء#uq 2W/nN~5lww7q0\&S,=|I9|qTLG31[RZv Bs:vV6hr0$brswU6m& —Ue*~yyxF׋:k:{h Q]Y4G~_~aZ(mdx<<[%3j'lTkEb"-Kj_ڬ3Ue 0 ~6B0nbۡO$E[WCM?fV'~ q#w"lIpVY1{W"UD-dkOtmgg^}>{MwC]r-YVXFw%~3}>2tvYċo' cn>i.:>{6Y-AOCCZTU{\76lDxjc(\5&꜁o`.hk_S?%ҹh"5qt0/pPIJ"F~p`>X.׽G[{#le$ v$ǣa+ # r*:J+|,syn'wrߩ}gհe1;јs/SS0~Br.󈩷ډ䗂b׫[ZV I@T#\4or  3C$BMJ#V<ԭ[ݫݔYv)V2DzmK%!]q)|2e l=Wte%N>06*miOn4VCšZ)ši[Оق2i|r ھ0YEz=&|%ù# Pz n_8MK6J):ADfdM 4oo'w6@gaa5AKħc')twYž~-O5Ɣ!R k;^kYڐ;q@ s~zV񎨇bG bMI.v pc턷8sE-oFS WB7\LcST^zTQ?{lq_uw:;u yCu}m\ݢ4&y<>;@2uIUT$/Q)ṩܹ^<* 9+Dƨ`*3/(2;s'Bk 4!.7bXtr$`?H9Ȱ\ysm!`m+_T+oR`Fsͳ_(V~2% aLάYT*Pk ]J2Gvu[sxx$p[ mIDK3'meBt UWAem&zXn }q1xQVRm"9T=ml(RWnΜ>Z[" 7ֻՇyI0($LvC٨N@vGKɯu[M-0*tļgaQ-ʉSցvE6jFaÌp\_f+_Vj\-&Ad Ϙ8̳tK?8o#,-z]:afm'AB@$ GI%/qpYGI\cDmyțCt:ޖd%gLX!re\a xyRs/LXOMs?B 8MN?Lp4+LPa6*ӆ P.Mg8[^8{XP,ʂ: ^,وV-=kꋟ)|㛛qҸѕ E(a#gTVi:ήuϬbD Ī*2mugJc@+Sήx(~t-ԁ1Tʪ4ޘ#*Wa q8;;Hoeâ6h﷼V@wpDc_;XİEg&)`璺'yElJϫ}&%#1W e9% )Y  q+rpfZRw جӐ[h2e̘6ӅHISשwA cpT`O؋27X1O`DJY煣aؾ)U6Ɂ*yJNa{1d(-h2#9Er Y p p̘㩨J(u?Q\Y3# 5n0gͭ8ϗS!J<+ҡx&RYkıQVYE<*~ Ϊq ^iJ逼:/xզ&꼏)U*q˘KIz,q;ôySm~DTq2ʿ?U/2o~.x}#C_8((1RN3^{SJ3 t/Sԉ|Ts<ɝչD!3rVHm10Q{ȂY9Ȱ\GLUh蒖˦g}pz -g89)',s]OF ksϾ* IX űJͽ V k0[R餁a)E¯1tv2u݉~̶i'}!L$[K$ % f%$tq3%i\qGּN lGI۝ J@#5jt&JT30vswz|/@]WVQ7 #~]=% sb| 8CVyژU{<Z%OOcz ui0l 0ń b^R=TЙYYYL5c4ME{3xc1֐Ǽy㩐k9S Ң|@ㄵ9EX=eGNԣh"- D\3IC>6ֵ& i$>%(ǧDy|"H(k8iIE>qzHczek>E*}uĪ{w\v{wuݻR [qK梁q;ϭ3>UnI94zֲҚV|z5S7T)o `\uQ<*[q+nE=|R0#rӣo{%nP難L8d!ӣk8 G9|yE5{?:u7/ĩgk[[OliaɆ#-8:B}BN,ȿaX6VpeӻOw!!vgֽ5k4pHSpFͨz6(3jڠ4DB': OaA%3jР4dɴO JBΐ8QDZ^:Lz_']/ UݠQ/ BBKI ow|"#Z] &ЊjTڏ?CDN]LTDU* F:oXHiaA,ԯ$+>jشS0ޭT^rɸYs"-RX)WdBxs4 (ȷs)ʭH jSy/,RYɖiC?=ioGЗE6 ]Ueb؈ɧ@SFאTgZav$ۆmd׻z/,}9Țqf8*d38K-V%:Z1\N`um05j5ѭthco.-'c*hc +n4M:$Y/3JIiMē]EEYs4qpD+{ކI%=k^qn jƲWIY&lF|:jL  `X/]K(z=;aL,tLTLo&G)kUON2Pk6*W1,j;O>RpOb Z#(4@ C!x;IF1'ɨl>kcJ3ʔ\;؅n'rVzо,X٧ԃ*~SE&Ayoěw>0H["&Ԑ>,{ϯl<$M5 π՟wdìn\%} 9A98GƟ4||32.{ N(nU<"X"eSdٝ4`4)SNqJ"I1dBTExOKY jd1rxș+gBS I BGT—7`s%CfMo:Z"ӬzϷ:,БM%P-͘_Vvl;xbK].t I4V [%g2;O&4H :lF*9j:H14rKR*}Lo,Pv`aۥbʹiJ~Ufցǘ YfGq̗蘕Y .ּNHI1hud7im jm2o |Ԥdo9h]P" %1"}+:3ہ֊66^\eEeC$Ǹ:oQ %QCBGʰmI*%W]R~Kic>8(>4x^//f7ڄ1[ B[3ͮ} &Oyfp|'\|K_CzA>cMyiaxwNB@fR!jɅtuGJ;/MikXNJ?Jsz"@J"]VQ|]XAB݇i:pEu^nO(,E,i9!>_1uad#Sv"[d^li9(nlQKjeH,EղrT-F˪8k6>Xd^MY˙{˺zkxX >}K,גq:=LǟՎf'C-E*αث*{$֧݋a}4p0>]}/1Z{9K^g҇7p .hJ}(J.hOb#,ѧ הYL>OD-%Ɵ\K)#E_P}';M2JQyn tzs@Bmz뀀6 )0EAW@) n|k4㫋/7a>zqu*B>Y~%iB}tX-rظY\ꌯ+vh3Q]X`CŹb;tjCJLK%7\4kH5.nZQ`wڠMT#,hb ٸKK#8c$ڵf>ON34[7o9W8[7ױ5jS|;p7ԖF́orꆲ`|ܰ8\4D`vs`Q@JBEp(%} TWHܭ/8&΍7!ZUY: NGRgPVRE,|h8u"6;jt˯Cn]Τ aU{C&BB?BaHDQ"&Er'M$7 \099nD@Ck8-\jXvωϧ5,ۡ^D>hr2Ttq<]PfK؆+s`E??``e͆- b!ɤB *+ָ V5YgmXCڱkXvhI٢xFY?fqYN9m9+k'\}|Pm`Vh"7_݊'eC34AH2ҵ͐ڙCΕu#G]mYIޔJ%T]u)X&gˡkyZƐ^މI͓"'E4O9{@] ).x΀9FD)cH"Z# t"]-0hVcJ򹼐m7aԊ5 ;Jr#-Ky1[ʔT%oX<2e{|m󍸈l`Z֮lLWUPUSgh5rI5y1&66`y1EZt$c{, ȥr [:ƸR$c#ڭ; ڙun AYP UYX+cQ0K$jH%l[pN( ǀ3ZFCyMJSp9xNxs'ٞg{RI7^l Z+Nř3=N Oxek(CH.d}tQ 6xPIo(|1fk&yAo, Ԏ}S9e[ԧ jH$NsH!yIx^eSQb%ZAZF҆3\0 8~ QԚTt&/ `P`}V!OH AI#҉(nWxO2$CcLJbԈvĨ35䏒 kׁinw{ݼ$FDq!&JJ{+֘^JMizsǰ:F Zq2p4bͺ'AY2;:͞ )g >T !F7SH18uL9'k7JsPL8 aj25֜u:~7W瘹d&AJ&X2uyAP+uV$Xy'm\t-֕;'s?".bWЅBҽ&)5Oxehh9^ɼd,;iDH_@ %J\̌,(Pt =:{OdB`BFUmV/JnOa>mֳo}5)ViӃxsnW*sňD_~G/-3ZMGꭟ.uq}|$Cz|YJfxws~Na֢j97ҁM?̌Ǖĉ1E38[S dP ftUZD)'وSe56z3.{&BлGTkL,v_e#bb\򰂟JBL{ɫ[L<ЌH7JF,^^<)&m-$OJ-$a^A+v3)ryɃ#4B(#=sG+x Gzz([в DLqǓQ2b<KmdA\6`] 1B,PR4b@AF6#YLHLjc^&MLd툐2u"a_ڦ"e7RPcwV1Sk̪lϘV"&d q^=NCQ]L_hʰPc2P ,LSQ1VAᆎ: (ܠKnQyZ&/o *J;T~ ]W$2^m$L|F7^7RƥK2[Z<EaQ5(,cmpJŌ ( ={.{a6R%ʐ*BD@yW>}w4-3]$%'QT +54 7;~ 3l rc'A6gu|Ju6|*-P?bOfHYn75'ݯ)} r@La:(K48z)er'-gvU_kiV8Xb za1L D~6"B)K%h%VD غS4<bo>XwN% w ;a;S4;U04) k#i 4F]ˡR1`VC{sTʉ TUuo~|ˑ']@Ot<t>@ 䮵4|*;ց//H>y$/'SAưrΪ^78pZlI('vp1ɡ> qnS\̯Llx +0ѳ5ZgGH^PP=C\r^x`$ єT[TU6lBjx$-1J(Ěٔ\ڥ>\^w,nbN/&4\ݕ&Ϩ]OiH^σ& Ԉ+8hRQ< F,So}~1G(نQ/b`?lקz ?>64@~l&R噠g,7*ӆxʥQ^HB0DKrIyT Gx=| K[<_NjEq6EȋfY, 4/MS ,мMD';/Mth<ѴؘJolqjN7{l1~ `l0z0e6Rbd ,A%Sct4m[ a˺n6F@@W7B{BW$=f]o?Xq!,1ɠXqpZHj?DfrJsg\zmZӸ;+N !MgcBDU=MiVβ4C$Bt@[DvITӆ1ON缁LaoI*e%X~ޮzq׽$Q_Ku#%@u6::/z]M!{uj)1 ʯx-dH3F~E慑jOsWOg_ ]77 ,|0uJywwSpo78qx3jsu,":hϬ $w1!:r\jsADT0L 9{cv _c'zi{[o,Ytk 8uqkG>n9^ s,VS] Ǥ$q:T+)-2PY˕RP!LdP#F*@p$8ynOFLbmi͕6KNxd$Luï B5Ej^}RH i.RBRGwޫ!sq\v\jpڎo!g5.Br4Fa"S4Q:CH)]~8|c$WiTA!5X9J1K&9<ar)NN Nd52Mu,G*KX"A i:A$Xru~ J&ZGjܛ2~_69YGfp8bc1qakϐe [C) AL6P{%I_()S ](CLVwXjpJ߂J#1Vj GjpMw;ʼn4؃ =| o,л{6tƃ&,?:_f "a3eVgGon<kȹ#(%c7Ѡ,V!W jeG3)GB5X 0`?Ød T ^?@DDhV/@1M7%?=Xg3DJ.)ՇoܿjƐ0el]P)iuzt#xLxar Ua.@$! 8+"~:w؞SrSM%E%Fqkx|z\k56ԥMV"%fB'RΤE!tAsNSkCYwALv)KCD=$p{HٍNC1Ĩ{AK?-]v.nsa֣k38F Q}->y{oNtX M:p{]eb 1.8!fIF6rvN mSj8ٖ&8Ƙ+rFk1(Qp8k1$W`2mrdetk= 9)XΜ:ͻ\Hh%MAՅ O\.052;h}wcE}-㜬cyڰ08 Ȥ(bDg) O$P}y]-\ 'kFc!-6q3 Jr ssS ǯ)jx5Wr~p5H+Z"ͭk ay0ogH+'#p;*x]ȱgPK53xzQKuÿ]^3* 9{{!J H0mcKi; _7v-VCj 7D `5´aH9 3*9QA6u 9ߖdy&Pc@s¬D^*[b%[en`@A "PSJ"`r`:49> 1 a`%IWG;)E5C*G1tZXZ"9} $J`LJXݬ3Pך3.@yX/oc8wQ~݉|C3m2^9ȘHS@9r@y(X( Q"8j̀l13g0ϫɘ.݅Irt.1]bZĦ51 ͗߻?yRLhiGOfYYWK?϶0e#]ѻ}z$O+4ȚrfD>T;e`W)DgǹV& {vY(^eʌ B읺=.ۮIݓXMf*4Z N2ՠ/] ][xJ٧v\jYK61YQ| SXlm߳' 30&YoicҐ+ڻ6?dkW2ՠ3(|CUfguҰݽc2!˾. ݛZ v>oiv>oI|ޒ{ޖQ&6Nnj?-78?Rɼ\YǠMJDivFse'\'FRޗ`;ZO I/@ID7=Xg>ŭ˻ޙlg:2 UߕNõG-"Wi{ fx 8{yEYneۣF;eԝv`X%Vvi ,6>W? yUO ? ~RP( x6i~0hx$VS]a|yyB㭯QDMN*EG;?a38[G[k9<.58}͏n,8*5QpkNWr@ҲLF'I2o*{bԗ3t%:KHR "h96}Q,}#"({K~rxOuT1F̟x}Hh+:u>~XҊjh5Jy)[8c0pP'—xtP"{JrҸ S0ٔM2glrFدLah83p>:9n5/dlڪ>ʈّ7~ЊY 0fXOBkꨆXih*uѕXUI9h:1-F#KKw>Oluek' ilZqޠբA oJ*,?ipݛ0~/=-ۖwNiT&CVܤJfʓ62{6%YͮnR`@V:Qܾ}mK߭D{ΥKN^JDI$6l-|6lsiZ2Sr{\K%R*^ US(NƧp/$WVQ.Q>iz^ Z_)vN_G:<2{ҚT%*RR9hQ`F}Dl}j w RXOw?%Oz) J'M%jて S8߭n_}LQ??kR{3uQ:U{?cuUݻҐlҸEF][Z H!pĹcN6Q*h R`vFN8u/vߏ.i` ^{u `XpDہc?}:d(ShVp+V<QKULSXoǧvK(J׎v7#wo1}Y "wD ہUG]CZ-;oV0Jqld+u K?IJn¦s`k-5SA+ha=ڝ*ީJdt/ 9 $x2G(l^?qN.bC'—_sNfe8F[/SIt|a-Lomaخo~oݾ_S^:oeDT9 iaʒz1ϨL;C:Mb?g /( J=CVye.{!J ?Rɸ5{a|v[JK[RGnl e aQƾ\bA +έ-9n{@V܈Ķ;G*q -Y #B8՚H|8BV1dBp ΙDJh!IP`QUg΀Yhf,!ܣ PHf%eX Hϰ)a%T@i凂pk@o3mRN` %QAsW3;q̕k)ˠ,XBZl,`JJ sV20i"_VjY0~\`' 78cY=ӵO?8!ƟIU*3~.Ժie&BkArR%hiURG_r|=~a~L2 KyMs" _|M8VDN|F)~`rH8w}'Giss"H+.K_Z!GR0s.cx6G_%`!Ĺ뾯j75J߹5XUޔ{~\҇4J#K>Fd9Fps?hd>c9F㣍haGb\98n~|X9n-Ɗml+ņX룍 G̎66$xcCBƆi'D` 16LaDK!3 !3Fic,ypȘDƆDolH$>ؐHu:Pӣ ALR66}d!mEw˫r~1=|`/9eNhb4h-T%, UF0rf\LI3\yn-Dh,D(pv#m1x-푐`͝#g;| vٕ<h"yLϖrÝ K[u=YAn4/yDRu.TzeySN߹GK}+=Z ]&M%&Xs&'1De5t9t`ǫOSo)cՕ-\ߘMX 'b{ifcZIZmLZ3 \WYFkKyk0<>yJhKNΰ >AGW-Ý }g[9h')pVw{;tLQ dۺ9j"I%({?~=,iV jj8WF\H|4urEV ɷ85NJ`7}06q{FfٵbYt"6<,[NWLZϹHbP-axSp5VRyгxs7%q<_<- SU}ŪYV1 X Sdr=>krVv4U4AB ݄zT N3vne޼[zٚޭ 9rM)E)vǻIZiEp@AI}.pu@ʽyB5[r*S?#;MaNwKŠ>cvc(=LKnmhȑhNOz7P? A-I|Np̆RN[rޭ 9rM)UiTٲmOUHŠ>cv@d޼[z;ޭ 9rM)ԻUT>wKŠ>cvL f{n jz64U4A0Rw7pt^.W+`/L\9Oaͤ(_kJ>)&C1N3_= 7USF!uRŞ-f!GI:ş~Fz;T N3vnՄŞM#kz64U4E%O>Ψ):g,b262g9,V-ArڷΩU9#.gĵi a^HΞSZwk+9_9&gڴa Z޽94ge䬌6-zHgMK1c̖j dB)̬j g̬efMKDz1kRS̬i 2kJP̬j &ZSL26-A3ǬAx$3V-Aq7f!D4Zfڳ)FXߘ5Zfڰ%Xߘ50E23kYk0kYQ=V0kSi5EBfZƬ$YZ$e(3kYk(SoZ̬j J5cm̬:%qf2֮%pI5Ar̬k ?fMZ̬j 1kP̬i ̚:WZ I5QV-A5zHf dfڵ2Ygf-3kZFRYÈ羫Yk@1{̙Y+2k) ٍN3ֶ oŋ]Ϧ0~(8?pg&D"1Z! 'zR0fixs:7ոm=*3sكƓNJu{ڃHvՐ2 7?xUF\08j=j=j=j= DJg;'E@ -Q{fxVBI+TyX zs([Ze^}~a>Ac `EZ-Apu؝gfao*&WP^ŕ aJP#TI1ݴ#{13ytG+^{[-_ KN&a]Pİ R|{N΃yaբimFUӖ2ԙae`VP`x+8*`і- Z(h !U~v6(ÊӁЛxffdgCxlφDsN!L^හтr1Zm 崘+?7m_\/,iu4Vb?Ӊ5g0 *Ȧ+o7V<7ðWe5'Aa'sn=-_n}0v6.֜"e Mۙ_<6/ڷ~f}+S.uI(^ٴ@OFc^ٲϼN,)gvtqmY.~qlz_bg~]E~3Tq<8{1(e2}W}x:aEDyT5"ap6D+%+>oWz1xV,n K^9&_lX+9-vhl<l@J]̗wY6FHoMBIRi9+$rS42 G_ŚI˻|/g:Մ5,`7pUǦ{eD`؊ 6,y ba`1rD21zI`&@p+?q'"rmb<)F~aGPN#߉c'Co&g6vkR7N"_I]n`BQ/eYvL*A-wzYո7>*su/]lM;Swk~Hg+o`;⎦u|[p^<꾊vqL6]b,n!mDfTO RUss94 FUIa\!ck9ttn'|!OXrq0|Iu;44itp AxiR|Ywחpnb?].f& `SQ&kml]oGW~9,Ce@l_p8l$BO?l]HI+}߯zHJ×գ!)8ⰧWUUjrsUM p}x]5@eo/&7CF,/?5k.o a6p>OiQSM (߹^+|>Mpcn ^PMAR'hҙ631珹qːFZ ]'0>>Ƿ=4p31EkdύkD( o/;$ + d%C\JN~#Ҧlm PO!D- 95b":?Vyyu,BN'XIO!;~N/x0u7'2I qv C[En{x)I*ƨ4·K6C=<ԎJи,e:8F4U7#֑%ӹ&E 5^@&/`DZ툶1`RgNSLqOoք [Զ!"?IKpLE2 4\MCVpnuiX4QE-Pif=HR-QnE\o۾Umۗ?YQar.dEq LNN*yX5EyI>JM5JᨕIppfKmN]M!PL'EK#|?_ 5_3ȘKltռHW £޾I[)=m_o3l;[g9ye4Otk3[Pu0Kw;OAR۞tnv]GX:b~"7!L|>} qN싰8{.@lr+(Kvܭ'kE'GlNIѠhQ3 ޶z2]R~l0aaqWyׄM͈0-x8n RKi(L01*VD9#lmEXfuK0Q74eh(!;'RX?JN*Ln%6*FQ?H|p|_eV$U;su7APr߈w)Z@øzڵኩR89\L!;o?k@twwj鰳ֵ=[{.` WtEIMurTn7<l<9RXkTI_ap6ar3VI;(O(7E)-R[ e0^C1rAn emUGʐDE4v9艹E{qr~c(YWYZIs-N|> E-[>'.L^9 Ȏ|8_*u&8jOmvNk4'ƿ2\gHʧ*%ȊwIi:jKe(,ڀ 23GK-L 6Ζ;X?@m6WBɈ@Ԉ b:Q:S53>Pθ$`L׵6yvYh]Vv5(xLeHmXPrd\Cm_/O*^*X [4JAAKNΧv.ʔ Q`@Pz5R#4?^\23?4ѣiX{GZ='>&pFx4)͵:hx"d#֘ [f*S#Q^Rzӫ|ȋM4e =,xJ Hj+jNjFk'};SIyJDo/ӵ~-0\azR}w\"t%տ+ӾnÄJ+_|[[ɔLjjA&+ݵu^ <-EDmE<إt_M)t+Uma*^~Hi 0cY!mFV ²T(Xjeo8k1 co*vg0A} .*?[f1s;~ \𣊫T<J'}қq󆓒 ('#'5GcraSJ.^V Mbec!qxWuˀE _E _n($Q/,A(K^0ˉWRV3zRn۾um_d 7T]7epM e]TI]_$u}E]]廤<Fgt xB"XH˩R\nVkY/Hpm[vVW Ls,zNwט6KF50f@VC @]ɥp6&*yJk޹oJm e*ºJzZ8YZgCfX#2k%IE]T +՜ %q3))uSm\_.^}y'۟ȻIuݕ'2j&j>n6 gӛ|sy:yuvFM_>;~}j/x0u7'sW{ naq.u:vdk;mP+hÙ|`KVȾǚ"3.rȂZCBJ4(2:b9pi Sﹱ:S Y oq}9 RDJXG G~h)p>뮦[*jk.T$TUJE"KJ ֔hbZ+_tz)lWS8(4p҆s3ԆgBЭ3_#U@pX'QnR“2)a|@vNh9 "Rz#ZLGV܈ Q3M;GE:Gbt*K͢A%3U < QA=5<߽WFv[V*,1ܼ1Y >G>"T #*h.IyTKEڔHjvېBcf 1UcH LfWo5nj6,> ;x: "S̵ugӯ_qтd& n~Q{;qJH<^< m;Exl ER@sv0?9& =v k@2UYË r*…(C9pX/@M4+NS4v,3>T30R˴r6gg ^LK:+\'ȱPk]b:"!L{60Ī!.g!hvrRXhCLGjCB2FD |#hu8 F~$\63tmݘQp>EJaEq^FxsgyO0&_UxlB/eױ`qP8}_ cȖ\$TExk'<2?ƃ&=ȍfs!ńw@V $E K [# ۠w38=+GҋA q%Q׽c=;[,Im_dp/"N"Q'#$ Z|ܽ6 L* :JzuyuIq{K`ֲ3{,w9@h9º3h![ U}PK Hn3iJFZ9q_f4“ʳˏ ^~_2#|;v?[fsn![9W޹XO}CU*4;W1[yY^\$?CWꐴ]X&GV\K*]艅ǝyk $@$^ܠNA a I8)6ezYM6e;1O!p3 HC1d{xwW%nv{BjTs%(wb- 5v7?^&;S H2t+J5 T\8"aO|r}9Xd,~2yu2ig_8rm~M 1V tn>AF3攇 k?-h:;xZe.yT]z%,s+Kj_X9DkIB#-vS;{nĈN3hy_S]fn酨ڭ y"$SBǹ<[*1&mQ#ҏ[Ej6$䙋hEhLxXoBω N/ s{3|,/QҨ3'3yq1Ζי)ųwU"U^[|VJxԹV74ڗwI/%1ʨT[ڇn.mqm 䂕{j"M*c,]SėDl-%J]E>q޹XOF[ 8*x_͈ҝe9 P 4#jIS 5!Ovj!cY@5 D;%ۙjQ^+o KPJwUy۫!8 Csuuؠ!$_;";uJ5{\L f&M6XGn.!!\DSdyP}M>햊A>vۣ/.Y[zHֆQQ^[ԉF !\Dw)PmV>߭aHF"C m Ы0A싦WGҫGfSJĘ G2 a0w6:0]BJ_Vg"gFHK"yҭd5vͬeȅ?U) =gy,ѝkNEkjlYK迄˳|j tqb~/Ww~a3?\]=M%fɪ$)%ۺ2 (Ql١U(DJϓaKȾVK8)SaRZ) }lB%m>Xh9=Z 嬆++>Hl+\e9 3!Ęm7&y:蝷X8ER: , '/N;|p? 'wB6-]]A=)4KRȝUDULLA\sw'vP$Zh___?]L fm4y ?I VIh~2οk6{ ~(i; 63 3e ,RkuN(p>Mj.xF6 S %oPð-x$ 3!o OJ:',TYt_lu5$}o%X$r]~F]z7{qI \ eZi!vcg BD ?|-Pœ&c1 Rˆx2̕ޡݨUcX%36/%z?V`ہ،L.sɥ& `9YGrY}ѣȏ6OZa z m#6~G?Ѡ.qG1ӁYA3a饹i#RlPrGX$r髿͊g8nOsZ4'MB,ܕ܍rw"SЅB,N F!iQL`T1L/,NZ@Cr$~r9 2.1DtXs$^._C 1bkNral^._zw~fy;}x]k%1[^0w#ƇӘrL;iEZnm7/(moi M!mp4py\ǀbh:=.e5@XmF&F s9 :dJF Ƙay%k۪%c{Kv&:`o}O4^DJE%XЏ-KT:Fx>JNz\1OemѶ.% 15V[ڢ= Ty*\W@aZc0\RC?4Et{>nd1;*[@g1>j纞S0HKEg.@֓?*܁1o^FaU!]b|渳n]y~(0}BB D7[+/.H!xRu]ȏl"%y] 8FHCuRVRIDJI'|}%V-aBTݢZ*͵MT]Ԛm'I:3uljc\5ǰxIcvhry7#x$i-1" ~O?]C|b񭗃p̸󹹽M.>۝L$яԒqlvSR- i-aB*6v`o/g3gpcK4(+L`4[rN`ig~Cx4|qy|,<ñdaae s!T9BLRB%_lDTŵkN Rr!V2-B\_8v]zˍvՖK~7/r0R_{Lsycf2`BS33I*~' \X,P$31'NZd 5w\ ;Xƕg aA -+ "%z!\6G)0ƥC>@v_aԈdF0">ZjfV"4ܱ3[&☐H> TE %?0.8ґ1VZRf+(8IyC|r\VSýՄٌIrs%Vg0&F9J H뎎Re>|gVm'5[ ZEP& >dͽ5" bi2 l0JQT&s"HPbc$7)[wzlަBּ l9%L,{"k¹%zC]SIK7JQ2YZ;;&;H}PwY 0)eVLXT\A93X<\;LCV j[ Lg":8PA1K h$ft4UYhc+:9.g .v;%(R`0le3ɠ7\m@ׁf?)n:IݺoD?bC|$*9sl}* {:Q1"90&Z).x}wOƘ O:_?]ЯĄByEL$MVp,;tAw߿Y PT?&w' 5@>i|遨˛W'5 skbߌ¡5ŸV=\}2R J4\p$$'?^]PaZu f5G<S&z%ucRF1w:K7w%t%S'ftkPÙX*3Km|+D-i"fp,3A&bFm"ᢼ H lm͘,TLD!VZ<̙6qWRtP&fl1쌑88ϐM$f5i;4NGL"A|;NVa^XoߝX ~!9"T$*C.΂l8X36B#]ȅ`+4R R_ƳοBIlqLt\d^0sSUl1R m*fA8DJJ:S0jC 9@opj F' -mp3C)pl$( yr!0¶fLDr:KFʟ Ƀ߹'=r0 pq53P 1y ,&|Qr=: tV2tUYUHY$nT&SJ%Q_e93\bK+\*i&ԙ:f5͔`GAzl ]U#wu.1e^# FSiq*!fX-dKm),d Dԇ{E&uIod8y1-W,)CCLr1=),wKy(U`0 p'b&#fsj0`d@ 2"v^3.\@W_*@v'b&#,ASh d5| ,$H( --u)4rS}}.4/ހ\-T@m$Ir.8 Ԍa%΁eOj,RZߦJNAu8RW]O/.d ;_B =ioGEЗb{I<%P]X)ʉ=ؼ&-3]zz!y1l, 3}]3O%8( 4ӛYeNge =8ͧ.¬Io_X<aXiGh{~ >ZUbՕ W9\xC5SlU\jYa_lg?}jLT:_W ݑt$ax־}w ;ک+8Ew .bGH1V)5X+@2HW֪bJ\WԒc&)g|ņ-_1U|m23,6+H1֑?a3$k6B:c$2'\d,N ]ٺߐEGW8Ph>Hs~jÊ. h .<%u CTM,**p嫛Ί)=m+ҝNJ>_skNʰ3,1CrgoPDK3Z(Ͽ9mzEFxX $J9b tS\p\[n(bYyƔoU 2PȎV8%ϒ4/dfZiٍd^?~f`6.o7F9nq 9BEVdQnCW[;EWѫ5,lރ.}ߊ1^WJ߱{C|wGĩ#W(󐩱KMZK*]kc3#K1W'$!߹I)2(B3*qeX6Y6$ڴ+Cx>VBaxeA8dzJ+=k/FϯD>*;_@w+`L+K4H6+Igb30 y<^Od?D_ξTbz)ĞV+R ,eT>[wQcMy=2$(#2Bjl[Ϭx`@xs}UIloYܻg;oR>N67@{F*\I{t."" 7N!x@RbIm1`x=HԄ {[0fpYc 25 D#ǃǀ0xD$kOOKQHCx-}tU34qc43pzːYS3Q< p* ?ABR>Zoy)xj?>Ȭfp5O? ֊p([58# >H"⊉ #FVH\=<ᚩۨ?3gcDV4#l]cd%w!ӯGƷ5DNC%w ƻq?.+уR!-(k û?Bb 'RUk7JS2|Y)#M:}<عݜ(3|N0r,)H1Xx/a=VXbAq1v5T(蘢 ]pGȥfamo }(faDR)\zG8ncj(w5cX+Y+8H%x 7C&v6KtMX y,wb-%@[-bH Քs Z; Wz$֛q>R+u# AZbEDN&Z0AsS;{K MCrZ#r~r9bt08I9{pd5qw= ]sn7>/ڧ|S;1~~B$Nފ1\FO[1r8 kk'/׿7:_#&GyL2 Ik^F6$N+qF #Aw[vdcwݗIB}fIB_@#"i!"jhkemݡfaj\Ef\EkQL= 5Ec~f jC؂SJ#ق?y|-8O>$;"j9am=ZRN;r|SFn峛[hLIzu 6 EtJߑcE8sz4V>%2Ѻ!!߹v)g֦tlDHAQmMk+!h$oһzD% V$y, CPw,~)H^$}yWMpI WtGgv0/q<0Guuy '$ R1[4X *F}`fPHiTW#8$}'IѹڙPZ'n߭'sJ+AN#m:[=AD !{V'l&~@IC&ť'05gMќSMEQOCfZx&(cOV|lB?քY 5il?W- 1EN~| tww-,h-xMJA[Gf7f&[hp_Eh\#E&ܛgHN<4T׈|ܹ/wS<"K%zeh/&njo_m   䘶;'/@"O%ܧg+@ߙ?GY8 ·0ޤ77pʀebP`,lʁhN+/*>0TԺT1ZW5Acj;n;f3uqG1A SU@_K<#e>ol]|~&68`"Ш\Zغ j! 9X(L 䴢d)=:~ӭnn>.r9'2tR^Xz†0%4,v #WHxe0ܻHƣc1?x}3t+7 ZӁPG :)x<,Gj-dmKIm,<*eG]qJp4KʊQ})?+uhj,ESVb_`&:,֭N= 偟(P$bu+"_ݏYWkI >jI2yUPŸԕq(wzJ|tl_%K!C@H*GG!N=$'ÌH$(@Q{>Pz+)Ȇ<Z#ASr<@aB5kW{$HJ$Zl}HG擨H',.//Tk ZL d'>h^Q@a,KȐhStݖI-"=ByI!8?m FHjbIL`MM1ҚMwH_pXLEda]K&cg[drȶN)ՒPS*e! k [-12k̎pTB)\Ͽav:MAuvEr/4 QIK ˟@el*:nv (, {l '0áN iRPG%ݯptJ*r4Gt%mN;3bLD-R&V&Z8=iHiUG x%T zشQy2E>~^@@ "o޿1ԭ#,v+GVJW "ԅ[inTe+BykZ!U-^3hzdyGƯ0^kk,N:6Va-8P*G!#m,ZG;&=o7"GK9 4\7gWf]d|IO\<̐2ʼT0ƴ^,h_Ҧ;/BhШ,oMm=R Vml. jl] [ňgmlȨb ZTC ^0I,Jŀyhよ~=UWg_y|?mʽD'9jx`BeeS"Ě-;Il*W _-s6z;Yi7_2RZ+*^K ji8Flvmd<#ghgZKW5PVe5J tБ3řw )⌠[4چ1S1k.i^hv~WśD{kɠEg3iC{%Ձ'ewZIE.~9|۞ŝ c)wl;^dxDC:D𻫫 lhdoGVH}Y@>q 1`rj(YDd|طXb$VSoQͳ& oA4APd.}Jsn;TS!be;͈ MNzt "a C>ovMyvjNX6KHSnrXzo|2/TDK!~SXɾ{s  A۟Y'_$*ڳwbH`AFd0ݪ67ŭ<@;"^,U2ަs( 2qVkwCntweDFhߝRPjEHx@ \!m֏^e}C㝇XYP`s(B\elL+*]ܹ?P V' Jɟ#q٧;l|)~H?Rc32ii4>p+4N~*fɫ%bY۫X8 VRGO7 GKaQUg kt:`c䇻Aw.eR:$yp'w~5j=hQlcތ9֠/瓋xwKdJiy?Y2\dfpچ-Ł$f \ M0 y^16FE<Ȗҁt@hb--gɷ%r|Y۷\r["tBү[I=59  Pe$kʴ\rHڵFb@3SIjVd]!1֌V9cJQ4%rgD3W/599=y{:_U͔asyTt?݊D@(I/fO,IONj!J1U2HY"4Q#+r+uN{x|TӱjW䠜dt7p[GfPŀdٰ,S>rT0KN#N'!sf0.5ʒBN_2Fa8O{7{{{{>TB1 Frg Qj/JY@$35a?))gTW:eA`+2|g&_thLg)˙N<#YsSDd21 B핤3԰;4Aڐ7Yƅ \MD<4??y~t'%d0qJ #F;&*tB4@ɠLͤv*pjKɃ" > ԅ aӓ1Wg5h׿= s??Y!+ywQkyqkn?@+ªvh@_$;[Z*)eßS-K,N&~WBD˘\y)0E*Y{2Y4O{dnxJTS h$q'h= &ժa"Z"4kb2D//|]ɭ x4W~1H =_ j: Gnm Of_b%ދtbln ǐ^'!]b߳aӻ3S'j]P0x7'A'SQW#!TL㧛Bj(G.exiu jʃiH5VǠ%D A#Kx@/EBxSȑց6:Ӹ:SQkzk1FjU>D( ӊ̅.ThL%7T4dSPD{p.hSVm5QӤ&x*^!$^(.TeG7{%R02zJ GJJcW)nbJVQōn4HAdC6g(7R7A0: uySÔ&x!oLrIbG [ %ἣgl_>3C\pǡJ*pdŨ~GlH=0`lx=$y6sQ&;c(x\4j3_p҅52Dڱ=ap"w43HSi%F IQ9y&ܛ(sC?(شdlk`P ɹA25/(өu6iYC[c1빿+#"uOMM4,I_yɧ_/b,49r2wO49k_/]|Oɦ;Mo/J@J[Rέ^h,#EtNhkiZЪu *)Jq"+j@y]4JC=N>yA?eĂ3'c x6fٔOK7zQAc% (n R0:?e5zu/;S{D<Erg*x]^fbShp]M{i%05M)eW"/;||lbnГۋ[BfCE*YY5|W Z#m*V|MsKߞԗZDƮ4Vj7 诔IObEw,[\|T) T&TTQjkh䧃-ZGsz~֣m~O;\L†1Ϗw}cڢ;rGn],:QǼY*EDobrV4ߨ)IiV}$sHQf>y Cjm|)Gr7yQJp3ʼnU&*x >yHy 44U$3^̯>U~AN^I]fLhWS0[{)WE(K**BY:ZlףLC<(>)YRgI -ҚgZ2^VW}vgrk$jI*{JD 13wm Ѝ_-#jEdĄv1q@ NҘ+ELo;PN~NM'z.6%zrDŒ %Rm[ OqQ@M;[qN8C bU#6R!9^!$QH$Bsd[8w+NV,P;^ڴzHKA>VG>w#.qsw*l%$=[z:Ϧ\pṔv$=Ƀ` T ---4C#+vRRM33e%6BDp%օ",$`'Ľu6uJ©0#D_Awb࠰A{_-fb j>ɦ]ca40OJYw(8+][ CՆ,רFLIB⧻7Il.y~w“"<7W]N9nzpan'9>r[*%~%H*:c,E>F#>\'eCg5F$X"q6b[T}ɹ*rwH00[;߿cc9A%K˝q"aW1B IhpKUgm} Sa6J SkꖚI֥*H[}0GkO66a[*s4N^BFV  85F`508A-̅2UZnSa6?&FiC+ǗUS?sS\i{\ȯ鋛" 3"!%*)l=%joKQgbQW:u&M)0f}IP„hPLah0`exíS5oI1~m{0>lmo~K dG J%jdmԸAPL69MsXoף`"\RZ`C8.7uQ|$0ZRs#glƺK0W-(P1ۡ ( TPEd)Zo,`-4+"nae @ Xk |wAJU*1*D9`G#IiFr+cʇ319\?)"1备Sc5]Ŕ%1.; j,I41A-ݓ+nǮV, /`h9')^zҵՎLWͮ qc6̹ߌo?ӏ SDŽT*c;cK!I]de ci'#0,re1`f VWnFVkOX~ Q֡4+W M- G, B= &0,dz8J ŋmbB%BI `hr w?Y8?#G@OW05AOP=._Ǡ$m "^>%CO{ko(F: 1Lg_HI` Nwnk&d8E=c#oˢ@ph&D~E\椂|8 4ﻍ_\]RxT`VBǶ$ /^FC{fcA-I1Yg6"T]kkb#P,<4&y>ICo78Bg+R~es( YvȎ1}㰝w#I_j+} %DZc8'kNU̓\G3fpiͶK40-a.K&|0.A@2Ɣe|M;1aDQ#HS&Ts` LŘHg ?1e\3c U>SЊxW8/9Ws/?=gPy#<צY3a9:UsX-uPRf2Z-T$Rꡔ:H1 s4,+lZbj#LձTMH…hE|X@pW*J3BQ3! 3l=W2~xariYmjߖMm~eω)w"\8h) "҇ Ilye}fMMi50pmwph0P#B@#F,X{J\( SY|9k;B?^l IYDR؜}zw9|=Nw\ߥDA)NSW5'(+~=Ns% zx`\q4qmI7f,WQ7HaK &( q-@TŎ1]k#2{3xK@0gu*O $zN>ˬSE4FIˉn4U> - ؈P$v0R.$+bn¥Ԍ[ÌZrXJ:ǸC0bUSZH$&݃WU ,N`΄h)dH ep/$FP)S9\1H~lM>oC#Pb;X&T!BG KC ZEI5<  B:6b8xt`*DDDru$,ZRS*ny'F]YaFU*l<)gR#eap.A*VWB3fC戴>}."hɺZxsp';CT+s:W̥䨓0IC&&ݧ), b+֨N:G Kٍf9c=lx2)í{2,AT8 sL,d#P| Glp"8l6 I7ւ&A'S|O/4K΄9ep v:xIX- _ vͣ!|$ S>@ǘip#OC f)@$zqcJP)up}5h^{3SnwkYzjmi/@FX8 $p H)P0x6B *^e(kTQ\eT rzcWmttD4:r :Lp5 *vtf/sZM|Jgvi-@} SCV30X&Xs?ˢI5 ~v'7Nߍ} Hq穾Uj a@CP %4yS Оo[rRB`"56V -_ ↛aJ2'74g.FVD("pBXٓq'4QlK F?Y2|AQB 7B L+"|?q\y 7})7ì}z%_fm(nnoyǀcHrF_6(1dzۻT{}x@[yH%JMn/F' "1_N0Tw*j'W盧ם{{=Vh!I_Ў;;N`<{79]狞"*ߗNKܸ- k]Gvq6|sL%◿矯k~.f|7^| ~1>4~{4L4$~ךOԮf@s#LG=R/RB*d"NS cf窴EHp0^#wKJnj]E cSJZa,ks8n+)jvs. 3w/sSu̜nG Ԁ8vr: bo،ʶp"GId)?k"t>u4hDшFlYpqq c(q*G^ ()цzV<)=2g@Y?s?\Ps:7"p&&Ȩ^\H+Z\Hz8EFMF'-2ZIT"u?AD(:z UtRcRg.crr'6SQ eC$W<7I*s@eQZ,8(U0qa[Y]Ji'Gsʊ~>S첗j\!e_H+q# )FJA7޵ƕ#"%X`K,Q^vEd0ݗL j+%$O%y\Z6 ]N*V-V釹+xwv{quwJMj"!\Mdȍќ3.$yûkVWֲ| ddVghuh1㥃Id0NFיč2ұQǬ q{~F {^QC j{F]( wkQQP6Bݨ(&PhU]BXc`m8ʩBsFJF"S.BL*ĸމq40x8/+S$/d87W7tOl, 'gדMW&%eNCe wt'|oW5g~psyOgӌgqA6h,/Y;[}b^[D_ԖLJRdrW<. +rr 9yy/~[}5E$S p.&Hr9N`8dNĉ%g`M^h=@,Zs$KI<YP{/ @?Ȣ4O9~4__yy"q=4QW'Ηf $R94N0rQ3򻒓93:Z `E #y%FsZ=-p=tbg]]_K} 0W[ٯJĂBp,iN}j5Pvo}#nZHwv#nAsHq譸UH˪n jW.U]dAd8PzSJ _'A;:rbG TlNzL_7"/5Zt+ԫgy_g]͐UC'p b "rɆ&<pⓐUP{ 'W>w"e3J_1(,Z18[=3c^zb k;jef_Zpz'7{vcPV^d#H)dqeyeҎ;w' ;, #~G.JF;粕tکv~]GF4EY]iZ˃Џ`OTALNNI@/9C0¡V$r@? l# j)f&"$ Ȕt$]LzUn3ɢ0^kaSN!uZbV5<.9Uhb;ΜvhYi[̈́[Ee[)y͌Eԋ:]:4b whsfYE1J@FQka;ΈZec-! KԤ.(,DI<7ʹ) LN8,tYt!;fWD{j_ʺL@ ]XfP^Tf5eJpr;Fn{~7 W:@>k;:a+Tš $;+ .,)X/K V=rܯ|&t VLnTvU[o<|9.xzt >bF.+:Rdk>+--f^`ֈnt5#]inwUa=يA+vqxJa~?nK}E\U\DQOtKq~st3ro.KՏ=?_λYjeѢˢEE.ZxAs B1 xS^RX4IL.q-dQ(KU"tMxdYS΋gV )jQ `=52~ lP rܙ ۸a{X=1|#~R*̺(>"Vwם_3&y\XO:bƚuh\^];QdW-Ǣ,'g}Z6Ag심Sbٻ{:o4ǿg@6qHxc[f#P3tJÚcPxs乩R}t7W/ /ZTVw/X{KI'=joÉr$'gh`' 4[|̴7=NԂ V׌5nӪ">Jo1(zôQ,^LEUGXfa&YWfhK2{TfP"TX𱛹fZvIyֵN ↎6*x@k퇹5>9UqK/'(9N'7xu;nhwK*lx*=̖{>C/vc(F(*qa8\XdoH֓usdZqYó}u5 C :YJnX)}}ʖPs8Fׅt`;[ Ao L%%֙( `c=} Frrth[4SrKj;md,(g0>ӆT32v"T Q~Y҆m8) (6{`7i6FNAX2H̭EBny'dD57ΕmR'BO!eX{oUA`BTjZAVuA;ai%^穘"\ o#tFKH&IbwZ]e!=Jn][4I(&3~8m3QM r܆JX C}<.'0Ț!>5 AGظc ߂6j歷1RLXk33%Fpa5#T2lCӊ? kAMU1ioS+@|EoG쒆yyZze\JE-5SfLN&%堇s2]fш)#pȚ_{=AFZP'XuƉv{ ebloJ],*wYTrͼ{HLnG,r-GYF%UVq E9'G)2RKUE:l(͘ l"f,5A2(h [uB Ѡ 2ِք7 7!!2V-z0dY8o[N Bl½*iWSf!b.2h}JcgZ",n^p4;H#~v5]0RZelZZjZdlʤ+;] tmp)hki.6`|QٹbSm$wR|ig>KW+GK|AT=U!o1G>&YrI۠"`;-ڌ wV{F!60Zk}~bpRW̙xr ̒tk8,& c"1$5%?{ȭj_Ry hэ:O'$TNUtJŖ-[ZIu:= CRp.%{Wh74@׵V2̝ǭ2薌Ld`w<'bNŧzn?%53m]VLೆv~eۨ4#lN*0[h:+?.hcTʏmdbYa{gV"W}%o ϿP3WdwSټݹRCv~f ϋ7$ ԃz6 ƃC$W!9N}w4?uPzLnx>_dޤ=xpDLJ<JdcLRg`&qaVY\h,!}Mp%kf4@p*#~ Z:2r2z<qʪ__~G/`\Z$5adчկ(>bU,,jc`C++v=E: ~dK7K=cxNQVjҳj7TѪ-:{^~~uݝXWӰ|TZ3ŏwB݅ suSΠNx#_Z|26ֺ >T8+BJ} b.*62(KbQ~#b/pWi{i7n<.OL{> :pү߹n(Y)n\f PidoQa)S@ٸ\ha otrrPI[V)8N;ɵAiShG.F@08Ӥhf[~ny&`z=y,Ѡ9S ک)o1!D\]J *-bBY.4)_Ρn;OA,gUC쨪L0U:6Ɖb܊l*eSn"ߊvӝMk˒ejlw+7}DxU0!ځueL"ȇkRxC,A4M%`'fy\wb5 y/=if;jUU)Qx*:bel,ey%^ȅtM..J}<Րŏ*!8'K䔒N89Vv#RLҔW"yprpXydoS]ɭΠ3n:K1#+_w{hw{F4Xin;+:9nZ[dEh K":xd UlK>^_ŏ?&c8yM)8VF:t@"M8eGP%uAu;$*%6V k(^t'N2% uf},Bkl'65 ua9:deq^MR߫]`tI~pF@(Z7x߉x2^ aS:j=OZ'Hl6 hҳ7<5khĕI2c'jkl6RGMS`utWfy3}q$xe]GvNӓp7t()nvU6VN?dSgLwK.&h|QY.wd?@M3u֒(g[nrF~zsw4s5{L)aEL֞d-NR`?sRQ88,ڇFy(FsQaN70fP`ovA(X Z^#sPl"c^#5m uOB!Frϴ'abv>J:f<^ 餑NaRTG*"x OSIVAԮjspUͧpUdzC%f<^J1J K<\CӯP7;'q8.Ngyvlo߃sjzߕzظ|Wb Q]Op<\S$i<:?`Thը]Y{huJ U Wz"2O) _.1jaEu8t!kz-qKlɜ2殓jN5*ӷ"!KG{xb5@Tb~eO{SIbt!7 RF)?RC_~A[;QCsE-&V9hʑ?'!eȟwQ:=G͜~^8]O:+o6ȍLZ2gl!q 8cf}iαc# N"B6NʚBV&_J"B @o_ąJWQ8)@+$UWUEEl+2WNuP]&"tdvK!k4 -ƴ@.bBBOĥ=;y+mJҋkCM̝gU㮗z-:&y=V6]|,}cuS[N6Kx𺚗zS> p|[^l=#[ZC^-5\S֖v:XàlXǧ'Wn2QSjm;Vg5mhX|my6R M?1#~Q0@iBF&=(P/Is/˺ɯq 9To0yυ lEe$ oغ +Jmw Z*%|6a\ Jo Oճ5Wz0q^e7O^݅`OR߉f):Q,L-bQk~𱔰[6va \XJ[!#--uJ ùBL εd4K)eߢ2?lwY3+\^-2zJ~0YJy) ̬Jb/=IDebXϦ ٢Bv'/~Vne5$TNDƜ,gB'3H%8v y{.z;L*96Su" ,}[M4-EzC]V-XM;c2[x^]ʰ`l49] "*ʹPE*7Dzd  /† OEjJ*'_> 0;Pn4*vW m I4~IvʙB|T (bTwT꩞[M(QR+F^>ծk z`ՃNY,-\zk%„Fo!4Xvv>|K̷w򖫻̠,W7: @}O777¯jX˜E`~;]|~Y6w*7v*'w MKdR[~S6 :!vSGI?QG)X'P( EGצ'TyQa;rd-ǩl 惝~l(j/5HR̄dҏg;J7/j}L&ײ<_x]B}2d/_6M|f|,+J{.ȧf]aueAtc(("hJuɺV Vܨ;Чo.>|S^Ta5OUcoW?]\oYG&3y{73ajOn] f- z}Ѹ-UZwg QH`|J`KFyuSQY;؅`VV/!LS{j͊/ ZZjwΒH (ISV n-A b[$;ޤa%omJ3AwzZ&#UpȤP#끦 'df3р,/ⰆFYTJ[I2_o7 ݤuȉfsrX5XG,=fxJ8"c7 }nKhIg5* kM_<ݽO~̏5~jSrVPqae$p׷NfZZnl - &︨(6HDpX749;#ymF(%}3 g$XSCuDU|HR>T:eb j'&6%c|o*sU"$V=4iL+eZi.+M3 X[)u *l FS"VdceVWjUq qn8M!fhZU^Agn&u8*"ԡӕiRTEMQDơ&*)`wN"[wMpz Nj 5 XTNܲt!WA&hUEYۊdd&XlxϢӀvȀz˱17&ڳ%LXKP$Z=Pi:)m4u ShIz} ԅJC'}m% kf(rqϣwi{n=n[f=}ᖿTIZn?[n?>=2w1PmLR1.8=kmHE˞j`fv 2B ͱ2#ɓ ߷ZQS-5;ʧ͈jM\ WE%>FS%!a̭7Ml@9Kĕ6P"<[ 2$9(`e։Y 9BlC19\ρx&Cc NÖ^K?&&v'R#mWh:_)<'T>]%^1a2tc3#ˡE^tg v(BƆsJ/2/ѧ̳76`~F&$W‰goVk'{-rF+뫻x1Lr ѹ|Mݦak|i_>:drivb#zN.h*V?VK=^5Bi]P]9 ;<|5O%&x>|Oy}ꖔ`kQs%F܀* TEv*dwm-G/Y|F(p?E* DS:E;`1333$τN[Y qe@2d1s.sxr|3I}VԨJ=5>? 1xp,D^1o!Ps+%Riz\4[EY.Dv@9W bHfConrY2*jn\ST84ԕN>ط%7sq wk?qi7f%dLe pB1d`R;t1m&cLԡ;ʐwKeȍȹM5 f̢롤FwU 7SK+?Ǿ  9n1ASK7H*ɳ)B54tGA">[O9OlhQTr.:%H-uό$?~1O@Dƴ)[ A⣸P_ JwNe@g)߂V?i4U>MsBJEQ0s$P]OpY/ZS;nړƸIlY)q ~Qkm.9!'~u=g1ӻ3Aqs&á=g6ۨL9Q#i VZ,%G T8%he@$ˤ4f*B)< G+[:b^ŏg{s+$#A?7$r>(I* R*/Hk @DgYF/֮u/y7bgQ:upO AY;.J1Gc/P^_/N7GWM o>N2r/^R?ʹr'u+mR1L,%$ ˾]Z:f_.x;9 MNj*1@Ƕ0{! z 4TxEW ,|J4SMQ>4Ef@M(T`Vg[\&ڧ洽78ǰt,F]95h٥{r/z/xd{jxDcZޟpxbHeTډ4=Wڪ8JscH#uMY ;^sp8x85І/x@!Ԣ hEEj_Jp_{n+XzLnAUwp\z klajSPlHsm4mnG:ݸ%.!V v/s߳Ԅ";=g1ME4iڸ_?H}jeҎP{(V&r^&IbsHIH >,WЊu24,g2BۀDuY Cƪ==~u-wՃzբ̑(T\p(]Oi9DgCAXȳJɫu.(s}d9 2hv8}ٍ_7 먥y}9kWMv9ܺk`"|CtF0y/:N m-.k6+KD -w6^ϾܹfٞiMiipPxt= LEYQu$uE!4T_NM3{*A!3?ڐ&t;/Y$}p{+)vLvYі-뿩/n>6:9]g c=B~(zPmyUjt9Uj߻$t:,=2v`= A^9'hȅzv" A߀U` MJbO9𡬧$)djƻŌ蝮 IĹሠMHPbIxu3nِ#NAĩتbMd_T5i"<Լ)2b%aBkh5߄Gh~gٜst;빢DCǒ/mCS[JmJq{|ecCsuc]9ZKEQG:hDHGrk/ۮCÎ\4BBMex|:ʹi0tf@Nt&P-"!Qyf=K&%yrWnjSM~?ߏ~j?,7?ޥ2LiZK?/nA]*{6h_}zƕ߼GFߤHt4aw$ Sb=8&6]?h}zS%D]6a Ml$Mخ_#iPZ=:[ԁM@1١۶KM|hF6F %R,j-$)8 Lǔ%K *TRMn8ҁJLmJ`m7@Es)wgMu7[wu_'$ә@ʍ6f25KĮ 2a25)JDjS\9>V} ٜ)}#,1bĩH 6!(-3jRH*h h>Z*-הfi gFX)xWUJqts=*}BJ-Be*!'mYuT X*~ DЌ5Q֔slL0bc<ܷ+@ұaVd)pzN'$u:kuFAWBN\jo ~\fN: 1MvB$U(Q,2 TAѺV ' D"IaP*LDƊs¦, +IbL oQpkR4Fʼn;#OHh  Ԩ@mʓD;Ys P˔ R1{|5Zy}p1:_W%Jl|6&6^:Y-Pocq?$KαOl?=J6~pM57 w@?,К"5w?TF|L/~j1?>NgデL/pD9 #jt/WΦ˛i'C'Ǩm z][|zp-x҈g멱n"sE%Z*00xd=}M0HT<ƨ( H1&+%x_oUKZ(&10\ӃWX3wqwC]^զVT@CeIַwQ)܍9$5 "}%W1u0Mj O R칑?$&bvW\k Us ˵PKefXN1,s),3X5H+ٔwkwUŗ9''?Y|Y)Wj2~KkHVKgza&OZX޻*!Z2JT&$<41Z,Kx˥EDJcD3vBjJ j`SLgŽpBPZ}<}q=,siVVTW=%W%r{Oz~8 VnEݺh8^767av g<yu岳dzvbzv1=w]gg*Zs%:[=?46(WiYRRs%v?^4 J{z1JM9])DZtJ%xOKR>bfOaR(D Q1G"Z\ a$V?hd ҝjV/ ) LGFg8DS)kbԂ@G%1J.Ll q?6hiIiIlYeSlRm"<:#Df8$鸌^i+ˎϼ3R}[ZZI{D1VV-Yfyc Z|>T>T7Z)}Mŵ;֡ЊsMH0\^ U%_δo a3|Q×c˄,sSL0ҪjDmNj̍ %j1:ŋSh9^s5ٟ BBX5Ŭ iԗP:Y۲xi;xZ9紆ZdJi 5Tdʂ$D3P RXԃe-f3ʼnJA4Z5+LXs5T3hh蔊ZI T CcZ C> Q~ d Nzyƪc޷.B^hЧɾm|d磾|/d 3Ջ$&չnZ4v(޴\>źAMs1MP+ xf0HxGg7Cƾ2i;*jPĴSS~ D7E-n /j.MxqfNr]o.M-[|=NX<\C-Q3OߺųLYZ=!򴆚Zdʪ~ZCmy ,:PML@ n iχ7z9"IپG\$.*e@(tXL_GDZ (ׅL*jA cc,hEpWcQ=Vaှ 3TWa2أ2[MRBay!~ 'e;%62RC4Vߔcm"s u#뭉NdB}| [Y'˱5|@Ԅ)hl6|jQyрQHg6rPZ.1zf1ZuVh$>@8sאTg0NDL{ 2N2WI\!b-ijNuE=JNdB}!t(NEIfNn»!yߛMSgx8N͙Y'VsќS}&\ Ẏ2Ǎ~zΌS'TԠ[o2W( *#I}._u OC^W .ґ_>2gn{q:M6Y?(ɼpf~#WtЅ6@W nO!Ћ)$b.SONһU JL-~R%~L.}:W(ˠ*G3zZXlV\v՛4P?ga{ig?|\Dl,W"/B*eK\]u dRTdhFb4]Kw[0q&L s>zRD!:;Oˀӗ4OIf[I5^8\o.ke}e W9>kfPw|("H۪0Xp#k&Pwx_e#FF'- ] tvGr&Sm˜ #mɠ *9}Pn]-G\W`҄7^O^nua뻊Kq\X\!D0EUE샙ldg9w8[=bR<,RB! (aM\ H @z+$a>hAXa թKeW%GmG8@-'b7VX;)/ŒNtu7_8 `z*ЋkOvu(`Hv'#Tw=6&}ٶtf8(Q վNODžc<|9;q*=F zHh'jh'S/QsOV1V= -jrVЋzz™^+2Q}Z'ժʜ]6BdDukƶ|Ӫn8-F:98-YPV坰ZiվkZju3S:Кc #PPa \X(L` Yeb)bjC"nfV@#IIV1K2#ޥT BDY EDE0(u! iI w5Z8DgrYpaBJa%E(5rH”ҚHkFQ@v}vPS0 :) '-b6 c\@njG1\쒷^69 ϫkk PI,l22vBJC?"80 LQɜ xԞ*0T}x)E#`dž*HCBa LAx!͂B A=~Fv]>]b1'ۂCblabf2t-0l#dNY9 ĞtpD8b bLuY0ecO){*5F("iRJs gNu3< -gg@\V'JӦ[$Fi Ue}.FQP3SdLJ)א)f?w@vM񢲕]E?eFu ~&Ч4(CE\.)>ۧ'}Q9ΫOADm>³XmyW7GpJj2qJ[{oCo]oh0=`׷&n,Yl=Ս14LZQ"K]kqWczPB6;H74''4%p[sH&md撿V wK | ҡXU{W)]wU{ՓXa+&FThtPFĀ4ObiF'4}z @skQn&kNX87XpY=sGxu^u6Vꑑ?t6XւB(43"%1`&(;UTwT3}Aoz9\cj>58ξ U=XVu K>~[\K($Y\}rūi(&*{s{'Ѣd{ӫ>xlěERpQCعߙa1 cؒ$#}_1Ix&][o+-N?%ppv7%dm!|%{FRϝ6=4=WbRAO/}qKW-׌Y#Dr#ӨhNj._ɽ?T2Ah%pUY9k]!ju31MyBް X(  pD e6´݋̓ov`D”3>V~&Oh 7\"~;Y6\΅my. E8!>'sũ \`+{|G G\ұPuo f+^>i%)&|f`d7NNdί̠F@2 'US}kZ(O|O n5E& c{eqk30|d3NzTبpΡ?+nQ^|Пn|`_yM_ L{g~ ^DkFhQT%b|?~>}ʎ Xm3*;ژ4䀳ƣWs3.;ah)ve`P7NN@sbYBoVL!"5kUzdz-'svc*Û򯺆aox]}'T^/mG+r @bX*ANI˼ˆ k 75*-X󕹒mw(V?cfdLF=v M6y^V:*0Q8X'K}}O+ku8_t|nXJ4CO c!/DSl ơǢ1wK tRۈncγ3\̻/MxRJp݆|V>r eAҡ$Dآ6:zijࣨ!Yp@~'׾(g[4𬀮`> 6A2Y#}CSmfHOF_A+y'9+8ߞ`474K%>$H՜N}?&{vW>m"L ;8*< -uyOF I0 >MH(ȴ`%U91,2NP%7莮  v>X`Ut(ęv(A?d/(ҌL2vq<`>ou=ՃDFK() Y),f<{BщćͤJ$Xe="& c^6B0g_ͺq7?__kuyI~/~\re,Aֻ.Lde6o?tתvIʊ8Uv/XeOi8|Mc8^#^cK>tk(aTnZvV穋 13-.b QF*D:vds_9EbRaTWNY2kHP aeX"mE fQ m%z8kVHr͋ f7ݦ{FD=fC9# @ /Nvrb7.g.9Pz ;fXBYƵ$9nuMt0B-"@m bQBrvcAdx\jAZ>H$w G(,vH- pkZZCF ĴiouMމ#4E1k.INQl&s *,T4RH3NIf+X[ (Z7HsP@ b1"JDo6'lq*Q|\7O2 ùs5uV \ ch 4R↓kAč-" SM B0ջ18Qv|h`Idbݺp@0)2f[}N/IIE<Ҕѻ5:2{zLk5`BM#iZQK!]MuN> |-Ii3³pD=9|SBF f(z$v9|ӷ99|n)6%P*Nn0wK tRۈnN{)ҹ[rGCMh$(!SИ [PA"hL{Ļw]:A霉^5} (8e*j' ykgW)+Bd뉻j񖈎Hj,m~7Et2u PњKƭYkc˜?-(Y(>$NFZ1akI 5|Ap`TJiI9툲a!WlȚՂ p.EVB iAEL 80݈*FfMlëSm?$.-ޙ/WrZ>=k>/BA~z[IT<};Gr3CADTV!_l^f뫛oܧs]~\&~v~q{FJ ~ӷg Ls3*Rqa(xqsIէ Q6bAoC–*ג'8Wru`tw Ȃb%vQ+mLrɳAe~XcIXr!pP"|!.AoLC@ѳ \O[(~ iiadw8㏌ &R)"j(8(D㖋Q 5-ՍVRjOn0Huc ڪcUvYͅr|{-ffE]ge OT>}k8OK7@]M2݊455Ȕ0͍kڛR tS㒙KoKC&8W؞{ " M8 iK91 MC~qc"8W .B1I:WqBwٟa ݺ̙^R5BF ˘ǔb / exB^S76rS[*!Fw1t:]ҷkMx;%<ںJP3NE =KBMԧ}ͦ4}Tm^n)M=r| JݞPiz*~^d]Rt@2u]qt$);ȁ"8 `w {.B9!١Jt9> !vH8`jRi\`=6֪ 4T|wQNI5H-(@NE OU5H +N~}'nc/g<@s_|~g{"D210_EX(" *":>p_E A5}K jTeFQ ,&Awyjևj=4s!03q*8BFp|N#%TX__T.I5<ߧB]m(\r #5))Q;GXone91o]Yg~|~}eӯva'.]zw7y[QT- 7ښ!L* "Ḿ0r2jPK\[_|w_.fYjCEJ5;o-/=|ĮƏb0ՆNI J*ZD#NS 4h!WAϰ%nmdSV{`C&ݐT?rS S)tikB#%-RAR,jjiGJ)k"źQ:xx ^m4'0!e)ok)HZhԂ` G#cCX0]G ( ⸶njĭ5 Z ՒH+m[DܘDgbusT2zIf4(Z2)9cQ~Ư1qեǮG}m5ssArѿ:>͛|E~QΥy+zG61wgc?>-&YOYRJRU-/Oꛩh8pN1G(DyTLwUsu&TW>nj|f5gښǭ_Q%pp*?lz:IT: X<ݹ-wW-p"A:eMqL֪N 16x(;rdt*m:9~' OȤ;.}GCA!JZ˗5Aq~GZ' YO' w Bm| ޺DQy:>JljPCi%]XDL`''0LqfIuԒg]Z UJSjO6-y ج}6Zp6-p1ٔԔc 7?H+܆qmSB ~Lۊn [S B\iu[gSEN`tkC޸S:yfSJ7Z ]nx, Jg#\; 1S6(5FNc>4wC5 y^?p8Ij&kmV]]Х{\ת2,mGAdq!vN_&r*O4J# ewLB )p}[gO4ITʽW+N 0gW%] l P Я".C_V~8JC@ߐ J,Ի%[u8YMբğ]o?CpB2GE'd? 2Ї(*:0 Ǚ9g[Kr0w#*S+/R8Q ɣMgH\XAA:@[zYgbhY;ёrei7I ֜*9QCzkJ(vUXLT}'\rݻ۫]RD3E="K3Mњ֨KeݜQ?%[R QPFu&L((E$ɄuV*ЦHLun(2e%z< 3ϋ£Gq0{)}s3.܇WߔK -C`-RzxxΦ %(5oYu㎗hׯqy 2B1.tAI>?n;{xjMH+^.wPQqj9'*sa7TnնV^Fr%mɓpL@DKk&x[ޖ*iFq^DҔdgi #pLJ܄N+۪ĹGԘ\iK4% J6MxsІuͤ  fZŒX$0ADDqwZuv1s@M\rQG%9m*y&(ɔ,O JXexbеxܤX:[_Kb~À O{O,&K!i.nIfPb`rN- 6HlSTkbVT ~F^l#Z2k<9攩l=ywتA/}Ct0=NfKC) '>25.JGpGl_EcA(cн`Ǒ" #zEh̞[87;20 :w{_h.] K7/[ Q.]3uxH qg$RxY@c6S] YHG'E^\ @Hl D 7<ڦpԆ3޸=lF YXH-zEGQ1 2 ÄE=,̮.eT~I݀WY |))6^z;܂;ϊ $6уgt]}\ҲEK2N\>3*˂;s͐xkZRilwm2B'qBƭc"*&wVj: myV2K !U $Q4Ur Nlgʹ*3$j=lHƌ j :koT?wAazlM^_kdDP(P; U?}^g7t]VtI }|3ىYy'sƇS+|M PgwnA" '|>|z]_%kd2Owfh̬\fawX E{ \Cmh8Ѻ_)IL[.Wps~:>wy||OD{F_r൥k ~[MV~-HJ'z2k&<˓'ݞdQe҈J._) y_|svs!s T \= kTqrvQН]:{d9'c-A=Aدݯ9˒[1?YcOC[q< ;NYw:}%թ`4n @uN]ޙN!r"3BxYk 'd$7͋_o%C{do Ig\‡whOa!H1uP |栚%TP;X_<+ s s9R?o, ,r}-7VEBn{ἑpZ95%IXyGx$3?oi${H0rX =Hλeye8ڲyHeŪDd dI @k2+ D p`$)nms聕D;(%A=srj{@-!%:B8oڇ9`mc~>>G"K!yPoULa ?` ybқ%}?<;2*+~1vI*nsgO&zݵku^|M21CSnTHD&$' Q[maVHGTl(wqno#Ngף-JsӍ-f4~ k]Iv0$;X8ʦy|t5jfrYmAl˲7# l< TZ驹<LS΋DgJgL*f@H& s#UEMR+SζܦreSCj%$b 3hq0 L̵Hsz^9Pq8\C㵧O]?\|d}]ӣy1{?E)Ы??x->u's!TLcg׿!c6Z[7d]ݧ0)[3~: ??߻}wՍzNͩ\}{tw;T_H=ӝ$cKW8x)9 V2qgrZ+XWnj p/kmϷOz#ѺΕZ SR]|39c_9 ?#iᐛS 3캮b]Z& @ A/VO`ZBDn :yxwyidZ aC!,M2(QjQ1$Th# Vd'N nU4-mt Ae:ikjm!.n&1'\ck=]XS¬Xf0\@J(Ds*ssF2M y@CJ\>7X̣rWi8/fQqnpdžןнC|vkȗKkaoH.ÍOw2(cȐCl M!b˅z(BubHcD/֦0xmO:uXԓ zKzƓ?4@8|JkwjO4ZnQĭkCwO@ŭOwnhEܺE<\tr/ݰоtriaA .* ' ՊjM^\rẇ9ud^ԩ I<9֖j,"UUŲhl 4jT|~,ց&I-jDscY4GӸKA x4Fh}nFkSpCw!JML+TfZdbL8A=ĂLLVyPi"AZn?:m Ks>P kd2N:luʩAS1퟈=$' !;|?0ea dCӵycIlbMFo"2Xnz\>ya0g}';J TsVwh,[aL]m隒0̜(+RՀZ# JlNg g\;8{=3OIϨ$qe˴W б8pO9d<_8=T72?St3DiRR P03Z>dG̯,e ih(4{3DFՍb2Z ipGТ5k6` DsFya뮦 D!u 0q)TK`I*Jvk=f2"‚pf<߃8c)E+ Ut5m/x]q>뵫3M"؞`;y/Tn.*xW79Cjxе=N S^/oךF'a1cTI/8hN3&Ȕbx*U_MJ$3O^KɠutǠ f|'8ϱ/ͬ6r\V2I\lU9@ $C5[6Yd| x*tΛ˄#xEUY9[7do~T kŠϲ+4Jg>=uyrVfa.y/fsYHqs7_8{2ƻe*ǽ&_><-@uvQ91x N(b(,idA\67]O6j6z$vG\gǟ|dze!'9\"vrXMQք"&{⼗8^ IӟƤ8/|ǢLLW("Nȵ qB%xp?vze猁Oކ[ݏSy~ӄߝ"edj+}^'+̶R0*Ly#H,> `!Nܡ(r=J`@L/<4Q2јͤ@VGV|ALV/p( \qQqTiar$̧~aĝxj{Ԁ~#<w)&n(H)fהgly=l>ʓHR=|\2ȘT^bXaøi]ňQI)jEɐ˧Tr*ĈBR~ 4@2V!(EX=_ [X@ m )zr A)B!U4J@!T}]<)$EgP6]A@A*"hzm#6+P` wa3 HbQm0i v-ESLSմ55AkeK0C5 6h٩ec&쫂?R6!R4G S )4@p-@ TVBJ,${JX#e+vfv@DIuƅ=bep gi BQh' Dt]eިGJbؙ  h_77o& LV0#9""<cZё;Ȼ  sq`FbVT ;*)55\ !5րwP-P-gTاH@l GOŢ&8f&!4i(m !Rkb\ FrU+l8|5Hr@:DKlKjԈVL6@qYd e !H jpy oԪk>_-ծ~4~_/ŧ-¶~؜`k|iZBC~j\?1Gݚת{'UW6_4b*凇K Ysm"̷pwg(Ww@B"U{>x !& M@ dǿ>Zf?&$bzL,ztqRHĄcӄ"d$#zS!- 6)&N6&I6) Eo̚Κ,&`^NbL y&8y)}v'|xv\&kelK{%`;4yP.N6@jFwxyx(:R$-R]Q-_) wRshm̯׷}gQ$Wˋn10p5dEo.:ehח| {櫿Ԙecn1&P[ߗ}o3cvcfwכc#yߡoy\}\s dקa)룮]=,zܵ"w_NP/9ٌPmuT 8#O>DƔTJw`ݤ,úb2tBc&^NfC['AcJs _u>L[粣Yq/k˚c&pk&#45#3ÕĴ֔BP#Z;q HMnIWC,IԒ ~!Թ_jV• +˭yW\Xå;;a=\$ Yϝ=xR$Ǚ).3±尼'=)iǹ XW mSˆ3 }ON1ޟs&شήb'5*D>^'P>N4ԧ`ZzOӳ0Eu %ۆHه\ DEM fׯ(LV(R+ X 5^oj&Gy̗$ DB2$ {tH0]rS0%&YdY=#qe6L_9ۋ-'7s9f# s,2@H;&m!HC)V֢ZH[Îijp㲴} #tfitbw0u0 s%؝ѹP<`4 L4Zj3g9ix) g:JIY#yoz,~>iUp1_X7ۆzX;[￾|I-^WW;ھ2fmy7SwyBy'wyBxetgb [m0^rq]o4_Tt ǭ岰Wr7unޭݼ ]X8=!'' '" "nSc N懖 d05fY5 g2 _ψ1Gs8s菟0+$^I>nP^"}KD ٝw&?8s0JDZ@PP/ec+b{3~Bϕ-p<`@8o~^~;$ځCuH&WG:!혍|j_'$_2-'|EN+PJ?^l4F߳^˦j1S޻ Ӈ ]ɡ 6u"vî:1`Oc'X[9#|j;jBMCpG&9(RV Jף /ldb}YẼ0 acCOcpbD_/ܹ}}_9xvL/l'C'[:v$?v(@1e0΁bim=fF7=HvpS'xӚ`<1{Bj_W;R/UvtG͚b;uk|u,Vr (L*O,B#*|>"(Ŭ'_,7+߮6>LbAnJTE *! mHZWtD9nra@qTmѿh$a<6!N]|iE7tCpY~t^1 #)(ޣrG8)~X'ZˆQɉkHi- !tݡVP5]mudqDWJY'4vSCwemI 213`Y#${ +FL ` Ò<1}  BHy+#i" 1ʔH"(e T{I\x }kByKnypn# " 5C5z,fV!aBA3G!T6}2K̯G_7%\zoQ`z,M*$6I((rX8)MO==KԼ/j*^TE# ;ʢyc//;Xh4xn5/^MX3y7ahТUGhA(U7ӓ]MK#&Qfiej2q985!}~Xޚ-pXuWlp%5I]/v1O^e<ˋO ZAОw^& =]V7LS^P*tq|8"+@-&u~+DJ*fOZSf(UO\Q\0KAꬻ.1oaJҿu/[OX֜pn䍋f٢*+q9:Ϳ8MUß|xpyW *)cp2o3DAuyfPXssޫ _MLI0م x-OOYgr68Ń,=Oq9dYۿ8I(-e#5T2=yjt17 w8}}fx9_nF{4cR@ (WT{~ W ]ٻ.#dxF@}[PE{F%+M}{gGcruU#{O,y9M+ 8nѼx^y ?̝4]4 8b_{.rC @Ws bmןoG!y'E,i@KJ"VYJ()G# Hˋ4@BQ-mZaɫޭ[dXCUA)B!FJ$UL'^&/J2ULZ{WP3PkF}(HtKnHa xf\:\j%QSdL<Ax5HP!mF`C-pjpW[(B,6p"@uXw48'C[FԁKh0EJ_ĻaSWpʥnAE@I\""(iF8έ+Jb" p!($#)ע0( I}b3M<:)ޯ7l5CMG!!RVjhS(ʾ@ &q50btnk!~Ɂ"˶Hph%c$֜:oă!*i\@-J[h4)919W_ ̊љif7$N ڰ@րHB[>x$ !+ŧG6bI*h T) #.*Q!%W*F&0Uͮ5ՠW^MI.S0^or N: k8J߄N}5]=4 j64lbVN^swe &QFN$PJ' 4O(oT/`E6sIh ؽ!LtuyLKyȹ. 6#L`C99<}qo0+,[{P寏He7fSEK!O_F8x ""SX2mSy/e!8oK4D%@XP(L l`,BB *bk¼!A`&#O1FB{hIW`>%|'R:M)Rhi3N)YHa2r,H )%Kª pn$\n>ݚzn"xW fq!#֍Z݀Ia_GzR_j,5# _* U> 5_֛rcܨ%4X#N]/B4i4=*Gv1Oi$&xU+ sQ "7-.ăЄͯɈ_ HիY 3ӟp::3L1@o\Ynvۉ3~I!#QYj6Tq)A XωS30_!!<;iU(wiO1Q),/Jd8}ᅍp E\PB_vNLhKLApO +s (^ϕ)sxy['ŲU 7# B|]LGq+|f% DZ< 7<^KG5ALD "h[KٺU>7F1nAfp62W2޶X3./xjQq?=a꺆a\u -Xr|\\NZm^3%MQTQ8= 2A@]_ofc)y&f"T"lfg!_/NrͫOV1Tm%¬ PLB.y&xAxs>OwWH$ ?P[,5nٚGz<t꫇xOk~v‘-%d5 'nurnsq4 *r4GJ*։ ~GJ?t1ub :r4D"2bnr?[#Ukpy ju1SހuYf2BLjZP[[R'5Z +L)~1-1ͫ]SR~=%jѶV9 zmw:|W_;eE=ߪ!9nNW9=ܧ eX~f -6ߩfՍ:O8ݼ{xVx6$A,iUK h-g5@#u~! |%AɺSljV:c=I_a3zKXV 6 ufw*I?ΨMISFRk4;(ǒHqr"Xߜ{$˶ζm둝RTکxgit׈R u]5ifS gTci*[p;wJ67Q _0t:^-`<-rq^Nk(8=Hf%9*k͗,:)q8KȗklDRE~eQ),gUsFlw ʆuzɎ'y\䇸DqR3>]w d+o|jQ3^-OXq wY[| :}I^#^??`GNbͭS "'@u@wmmJ~Y`qhR$6s =@5q,_F$[d%I%+$H,K*UEaKQGi;/%Oo$d!^[u4?HM[I7tkf.G57U,vR1b)ӼJM%] DJRb0.!W hL|V 4%9?1 m6Sx4^1@;;]p'Kf߁Xz+Jm>׶>D˭!OȵWzݟ,pf:[ފS܊s>5b9ɧQ˅[Q;NtEU!g FE0IHb㭩q1\1]`on*?|2 b ` sxko!OJ BrS(ў~*vb )bwi~d63XXb:rk71eA;w$X_"wƊFCFV\Tv@ l㓃LQK&8t`A2&6dRA_G]ھd#d*S:1HL5e2a8%{wv$E>d^%>KVdsoWIRHlă5k <,qh~0&5 x}JH^,õ-Sb6Lf_*zNk}(yB|u g4J}We=c9G!ABq3'RJ໴w9|Xb!Fd_.&GΠ w@zţӋ8ߕx^+o[sq@V9`Iq߯B 81Y%1Gh" 9X> PFb&f6>g"$)3%k%NF]Mr2|u!xf$7 c‘[o7tfq0-ikETSwήYN=wr]ry^ɰMmӲ*Z=pUЎ{Y}د}y q pmiM톖#e{Z~|8' ۋpߩypˌ)y٘tolY6L:vso>B<-Ps X{fST Hr:KA4 W3]UhIg?qqJKx dh3"]]8>9.1'G' ^K>q"ӧ;iv[`)"%^ 'sHbRnd;LP8|ӼE_8kݞA5ӢQth7VcN%5߹T UC MHM5=AG/xnYsO(`E+BcHfِDA/P#s%킽hxI Lj,XoYpJ* ^~>n2G'P,Z\T)GJZw2_y\@L%w,'BK拔,iLEP)xq7Ja@#,d$cRY.a\ MQ@Cd@‚__ p3*F[EAi8O규Vq[UpQYBx,%R1\EeR4}`Fo7fp~'>jnˌhD893* ;6!.o#ͅpY,4҃p&BDʪiް=!ZP-^5Dipgq֖B; Rc7S:KL0O(;TɁ61l_d_D \,MQ.k/̣dtrgr5Ђ8Q4:'jǚ!݊i?ݕzJڃd"Ԟ^lN *[1ӗXo}Y|{0Ll]^qqb;Ӹ,d R^DR`:z{,]hݒ93C yg vZ37>Ҭuxi{(f>L~P~Ͻ7H8/O6Hx4Z9?+mK WVJ3pX*V3&88ÒJe匕n<rځ'\<ѝR̂sb440Rꢼ! V @a: wFI-! EA[$]B8{cD܁u=g,d?#1W`hjis9pb 9u^сuȟkY_^nydv^(lxȸPޏR0_,gPbdBeM]R%>;X^%S5p?Zr1"]-WR^tι"]+(alSng_h?bC DkF2"@G[^5D|:|y>ǻZ'P {HɷV6E#'B: J9cks{k]n;5] L;;+i:x0ųMHRx{Qy@.u$~-ᴌJN[B$jfNEy-i>@6 o.Ed h@oFQKa)`~%X<9GOws1rk[ nO2_Cbd)ɾ+R`[Ch+䛀Jq⍨0J!oCicaBm3娖(fn@,,**#K:M7.֔`ht hsA5JX-Du "]QCs^ca(%9XY))S"ĸ=5qQMѣԗ[dQ-ZR-NUe?z"Qi.5-'z݉F6tLykl4{c,^j$#td$ڔ <rOTcK ;sf@փZqr?tŃo<Ө8fKb*Hm ,KN3CI.!;[TklpQ9^ꤊBkڨXn[F-Zz .UڛmK{6Ԏ0aIv|y3ț0cXc Y@!z`x!KlA=x]c .'שc%ejOrmh x1;zR}}^/"i8-HӻïyŪ'63 `՛.o y*KMHuj͗n}yFu>u{Aߌd8k_uBC޸܋ uՊ3;%eP?ڇ<ˑ(;$\O˕,yWLݣ l鍏yhʕΎqhNrcv!ϻTs0Wux!CtGxsz\i(xxqNTNrV?_v]XN_a24 o{6]fލh+,znP,^˷ kL+,C(@YOː/6Kn6Ny6ml{ \:6h+XJMdl=?jc{ȓͿFЍٱ aoGX'W;?~?J?+|8i~,O~~.'f bAiKIKUܣژD|j߹j"B-w>GU[3Jz?|{F qDs]y_$VVUOc1>W;9׉T܏rOszs%D -p=jZڪgOaas-6>GK[qc-ܨlC}sRJ(}P@x(%pňZR?ؕSVS{s]TK&t`QBgth Nwr2^{lv g0$:'cqq;Ӹ{l{1mPI0-RlP6}cj&p{s~:YfRt.'yvzKI^M Wʱ# {I ˹u܎jjYp~wbmtc:-]XXiׄb-l*F͜H8lL[p譄A|. 2^@ԁ5 TL cl)e  ˭EhZlfHЊ(#v8P먁E6=CF ):ƕf…R>` T*ES[$k>}fGk6\~pc}h5`AcjZ+϶@\L_ASMZ5,a!FBL٧+5#lŠ] Ue21Qv;.7a=(a=rbj(PcZPB-g0WQ荬\ypNc]vc4/lI}ǚ JR(r45XRXrxfmE'v)PhÇQ}h%;ghUVU*XlM1kLM˶c +CzZjF"Tu\u r`O: @TL^$UФb SP![In$m߆S͞ FpƱKt֙Sj"HjthyAɒqt<$=T:rh,$K#MHzXkt hP S٬w=6x`ԨTs{ M.K= sC i RͶA+U`2LcG,w4j*a"X}pA' %MgFPЪWշUn&^KNڲ|JQP]ݬ`STܮ*gƒZ'X{)mRI22bnGPmu+8}YaH%Ay Uf=-}NFNhG0CK]A%7ϟIr%. h{j߮SPNZr3EwSo_/և̄e,"'žJkuY!wRáJ{QT}#WwK` h5 +,MBNw͖w%A:Sɱ>:%$ LLŃԼ~4 te.Q%c_&R *?"k':&K`,+Y]IӹoTl kZ,)A;AV䠥VHm~:d$I[yzC:7(Ύs]L%%oE:ѥ7W㣝Ja3>q[ِ|hbgWi,켸HhXMH7&V&'uθf:&mM]J3$΋?{!U(cT=]IihS.I>m!TNh \ [E QZ9QH+{o) ow0֋?M0U+=bb_"NMN8U8<\iA zrܐZ׽:=i$X;ih{%߀JOWZaI_wLvuK}c&JWy/Le9e䭿OC!$-quǸO}@.KUfܝd1Z]' nkbdmjSdW?e'%KSOɘ(%GI@{ݜ% ^@鬴Gl}u˵sʡBw[pg@evv=vl⏍>֠8u>ZO;_j̣LګehN5}V;jB̞¨}n`L5 tRMSIu/ǺޮSt?<[1LҞ7\߿ ݟӛa]Nz}yiLן."dv?2Ap-ڸ 0o:6S^e-/xn,?diՄ힅<SKwV.N&\DGɔ5GGk7v,:nc1gwQKMOnSHW.dJ7k7rGbP1|G/p.OIn-jm $+5LV4ii&l5?UR3–]atMvۏ&)aضu-$B@!C6E 7&>Avar/home/core/zuul-output/logs/kubelet.log0000644000000000000000002023725215154271517017710 0ustar rootrootMar 11 13:22:31 crc systemd[1]: Starting Kubernetes Kubelet... Mar 11 13:22:31 crc restorecon[4690]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:31 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 11 13:22:32 crc restorecon[4690]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 11 13:22:32 crc restorecon[4690]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Mar 11 13:22:32 crc kubenswrapper[4923]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 11 13:22:32 crc kubenswrapper[4923]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 11 13:22:32 crc kubenswrapper[4923]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 11 13:22:32 crc kubenswrapper[4923]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 11 13:22:32 crc kubenswrapper[4923]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 11 13:22:32 crc kubenswrapper[4923]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.748408 4923 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760419 4923 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760470 4923 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760479 4923 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760487 4923 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760496 4923 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760505 4923 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760513 4923 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760523 4923 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760535 4923 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760545 4923 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760554 4923 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760562 4923 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760570 4923 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760578 4923 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760587 4923 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760597 4923 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760605 4923 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760614 4923 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760622 4923 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760631 4923 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760640 4923 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760648 4923 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760656 4923 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760664 4923 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760672 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760681 4923 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760690 4923 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760699 4923 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760707 4923 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760730 4923 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760739 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760748 4923 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760756 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760768 4923 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760778 4923 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760786 4923 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760796 4923 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760810 4923 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760818 4923 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760827 4923 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760835 4923 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760843 4923 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760852 4923 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760860 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760868 4923 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760876 4923 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760884 4923 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760892 4923 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760900 4923 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760910 4923 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760920 4923 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760933 4923 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760943 4923 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760952 4923 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760960 4923 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760968 4923 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760975 4923 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760983 4923 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760990 4923 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.760998 4923 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.761006 4923 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.761013 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.761021 4923 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.761029 4923 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.761036 4923 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.761044 4923 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.761052 4923 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.761060 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.761068 4923 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.761076 4923 feature_gate.go:330] unrecognized feature gate: Example Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.761084 4923 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761259 4923 flags.go:64] FLAG: --address="0.0.0.0" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761277 4923 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761330 4923 flags.go:64] FLAG: --anonymous-auth="true" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761368 4923 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761380 4923 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761389 4923 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761401 4923 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761412 4923 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761431 4923 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761441 4923 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761451 4923 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761474 4923 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761485 4923 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761494 4923 flags.go:64] FLAG: --cgroup-root="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761503 4923 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761512 4923 flags.go:64] FLAG: --client-ca-file="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761521 4923 flags.go:64] FLAG: --cloud-config="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761530 4923 flags.go:64] FLAG: --cloud-provider="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761539 4923 flags.go:64] FLAG: --cluster-dns="[]" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761561 4923 flags.go:64] FLAG: --cluster-domain="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761570 4923 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761579 4923 flags.go:64] FLAG: --config-dir="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761588 4923 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761597 4923 flags.go:64] FLAG: --container-log-max-files="5" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761609 4923 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761619 4923 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761628 4923 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761638 4923 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761647 4923 flags.go:64] FLAG: --contention-profiling="false" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761656 4923 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761665 4923 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761674 4923 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761683 4923 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761694 4923 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761703 4923 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761712 4923 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761721 4923 flags.go:64] FLAG: --enable-load-reader="false" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761730 4923 flags.go:64] FLAG: --enable-server="true" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761739 4923 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761756 4923 flags.go:64] FLAG: --event-burst="100" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761766 4923 flags.go:64] FLAG: --event-qps="50" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761775 4923 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761783 4923 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761793 4923 flags.go:64] FLAG: --eviction-hard="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761804 4923 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761813 4923 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761824 4923 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761845 4923 flags.go:64] FLAG: --eviction-soft="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761856 4923 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761865 4923 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761874 4923 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761883 4923 flags.go:64] FLAG: --experimental-mounter-path="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761891 4923 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761900 4923 flags.go:64] FLAG: --fail-swap-on="true" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761910 4923 flags.go:64] FLAG: --feature-gates="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761929 4923 flags.go:64] FLAG: --file-check-frequency="20s" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761939 4923 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761949 4923 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761958 4923 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761968 4923 flags.go:64] FLAG: --healthz-port="10248" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761978 4923 flags.go:64] FLAG: --help="false" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.761987 4923 flags.go:64] FLAG: --hostname-override="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762005 4923 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762014 4923 flags.go:64] FLAG: --http-check-frequency="20s" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762026 4923 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762037 4923 flags.go:64] FLAG: --image-credential-provider-config="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762048 4923 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762059 4923 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762071 4923 flags.go:64] FLAG: --image-service-endpoint="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762082 4923 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762092 4923 flags.go:64] FLAG: --kube-api-burst="100" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762101 4923 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762111 4923 flags.go:64] FLAG: --kube-api-qps="50" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762120 4923 flags.go:64] FLAG: --kube-reserved="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762129 4923 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762138 4923 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762147 4923 flags.go:64] FLAG: --kubelet-cgroups="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762156 4923 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762165 4923 flags.go:64] FLAG: --lock-file="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762173 4923 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762183 4923 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762192 4923 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762206 4923 flags.go:64] FLAG: --log-json-split-stream="false" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762229 4923 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762238 4923 flags.go:64] FLAG: --log-text-split-stream="false" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762247 4923 flags.go:64] FLAG: --logging-format="text" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762256 4923 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762266 4923 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762276 4923 flags.go:64] FLAG: --manifest-url="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762285 4923 flags.go:64] FLAG: --manifest-url-header="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762297 4923 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762306 4923 flags.go:64] FLAG: --max-open-files="1000000" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762317 4923 flags.go:64] FLAG: --max-pods="110" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762326 4923 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762338 4923 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762386 4923 flags.go:64] FLAG: --memory-manager-policy="None" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762395 4923 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762404 4923 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762413 4923 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762422 4923 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762441 4923 flags.go:64] FLAG: --node-status-max-images="50" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762451 4923 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762460 4923 flags.go:64] FLAG: --oom-score-adj="-999" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762469 4923 flags.go:64] FLAG: --pod-cidr="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762478 4923 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762491 4923 flags.go:64] FLAG: --pod-manifest-path="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762500 4923 flags.go:64] FLAG: --pod-max-pids="-1" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762509 4923 flags.go:64] FLAG: --pods-per-core="0" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762519 4923 flags.go:64] FLAG: --port="10250" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762528 4923 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762537 4923 flags.go:64] FLAG: --provider-id="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762546 4923 flags.go:64] FLAG: --qos-reserved="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762554 4923 flags.go:64] FLAG: --read-only-port="10255" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762564 4923 flags.go:64] FLAG: --register-node="true" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762573 4923 flags.go:64] FLAG: --register-schedulable="true" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762582 4923 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762597 4923 flags.go:64] FLAG: --registry-burst="10" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762605 4923 flags.go:64] FLAG: --registry-qps="5" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762614 4923 flags.go:64] FLAG: --reserved-cpus="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762637 4923 flags.go:64] FLAG: --reserved-memory="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762649 4923 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762659 4923 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762668 4923 flags.go:64] FLAG: --rotate-certificates="false" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762677 4923 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762686 4923 flags.go:64] FLAG: --runonce="false" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762695 4923 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762704 4923 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762717 4923 flags.go:64] FLAG: --seccomp-default="false" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762726 4923 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762734 4923 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762744 4923 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762754 4923 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762763 4923 flags.go:64] FLAG: --storage-driver-password="root" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762772 4923 flags.go:64] FLAG: --storage-driver-secure="false" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762781 4923 flags.go:64] FLAG: --storage-driver-table="stats" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762789 4923 flags.go:64] FLAG: --storage-driver-user="root" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762798 4923 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762808 4923 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762817 4923 flags.go:64] FLAG: --system-cgroups="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762826 4923 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762839 4923 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762848 4923 flags.go:64] FLAG: --tls-cert-file="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762857 4923 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762875 4923 flags.go:64] FLAG: --tls-min-version="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762884 4923 flags.go:64] FLAG: --tls-private-key-file="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762893 4923 flags.go:64] FLAG: --topology-manager-policy="none" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762903 4923 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762912 4923 flags.go:64] FLAG: --topology-manager-scope="container" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762922 4923 flags.go:64] FLAG: --v="2" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762934 4923 flags.go:64] FLAG: --version="false" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762945 4923 flags.go:64] FLAG: --vmodule="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762956 4923 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.762966 4923 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763173 4923 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763185 4923 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763207 4923 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763215 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763224 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763232 4923 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763244 4923 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763256 4923 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763265 4923 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763273 4923 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763285 4923 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763294 4923 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763304 4923 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763313 4923 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763321 4923 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763330 4923 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763366 4923 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763376 4923 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763384 4923 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763393 4923 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763402 4923 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763410 4923 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763418 4923 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763426 4923 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763434 4923 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763445 4923 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763455 4923 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763465 4923 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763474 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763485 4923 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763494 4923 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763503 4923 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763512 4923 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763521 4923 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763529 4923 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763537 4923 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763545 4923 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763554 4923 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763574 4923 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763591 4923 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763599 4923 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763607 4923 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763615 4923 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763623 4923 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763631 4923 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763638 4923 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763647 4923 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763657 4923 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763691 4923 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763701 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763710 4923 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763718 4923 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763726 4923 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763734 4923 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763742 4923 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763750 4923 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763758 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763767 4923 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763774 4923 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763782 4923 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763790 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763798 4923 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763807 4923 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763815 4923 feature_gate.go:330] unrecognized feature gate: Example Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763822 4923 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763831 4923 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763839 4923 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763847 4923 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763855 4923 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763863 4923 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.763871 4923 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.763900 4923 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.777211 4923 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.777266 4923 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777519 4923 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777542 4923 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777553 4923 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777563 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777571 4923 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777579 4923 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777587 4923 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777595 4923 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777603 4923 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777612 4923 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777622 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777632 4923 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777643 4923 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777652 4923 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777662 4923 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777672 4923 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777682 4923 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777692 4923 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777703 4923 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777713 4923 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777722 4923 feature_gate.go:330] unrecognized feature gate: Example Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777732 4923 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777745 4923 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777759 4923 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777770 4923 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777780 4923 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777790 4923 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777801 4923 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777811 4923 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777824 4923 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777834 4923 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777843 4923 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777851 4923 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777859 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777867 4923 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777877 4923 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777887 4923 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777895 4923 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777903 4923 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777911 4923 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777918 4923 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777926 4923 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777934 4923 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777942 4923 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777950 4923 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777957 4923 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777965 4923 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777973 4923 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777981 4923 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777989 4923 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.777997 4923 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778010 4923 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778020 4923 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778030 4923 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778041 4923 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778049 4923 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778058 4923 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778065 4923 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778073 4923 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778082 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778090 4923 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778097 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778105 4923 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778114 4923 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778122 4923 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778137 4923 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778146 4923 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778153 4923 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778161 4923 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778168 4923 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778176 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.778190 4923 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778459 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778472 4923 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778481 4923 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778489 4923 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778496 4923 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778504 4923 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778513 4923 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778520 4923 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778528 4923 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778536 4923 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778544 4923 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778552 4923 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778560 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778569 4923 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778579 4923 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778590 4923 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778598 4923 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778607 4923 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778616 4923 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778625 4923 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778633 4923 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778642 4923 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778651 4923 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778659 4923 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778669 4923 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778677 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778685 4923 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778695 4923 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778704 4923 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778713 4923 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778721 4923 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778729 4923 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778737 4923 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778749 4923 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778759 4923 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778769 4923 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778777 4923 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778786 4923 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778794 4923 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778803 4923 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778811 4923 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778819 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778827 4923 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778836 4923 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778844 4923 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778852 4923 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778860 4923 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778868 4923 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778875 4923 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778886 4923 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778895 4923 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778905 4923 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778913 4923 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778921 4923 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778930 4923 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778939 4923 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778948 4923 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778956 4923 feature_gate.go:330] unrecognized feature gate: Example Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778964 4923 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778972 4923 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778980 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778988 4923 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.778996 4923 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.779004 4923 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.779012 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.779021 4923 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.779031 4923 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.779040 4923 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.779048 4923 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.779057 4923 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.779065 4923 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.779076 4923 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.779337 4923 server.go:940] "Client rotation is on, will bootstrap in background" Mar 11 13:22:32 crc kubenswrapper[4923]: E0311 13:22:32.784939 4923 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.789802 4923 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.789935 4923 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.793253 4923 server.go:997] "Starting client certificate rotation" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.793307 4923 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.793684 4923 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.823071 4923 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 11 13:22:32 crc kubenswrapper[4923]: E0311 13:22:32.824976 4923 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.111:6443: connect: connection refused" logger="UnhandledError" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.830224 4923 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.844483 4923 log.go:25] "Validated CRI v1 runtime API" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.890497 4923 log.go:25] "Validated CRI v1 image API" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.892800 4923 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.898237 4923 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-03-11-13-18-28-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.898308 4923 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.917863 4923 manager.go:217] Machine: {Timestamp:2026-03-11 13:22:32.915846234 +0000 UTC m=+0.597747988 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:cf7f6afa-08d5-4a6d-91e6-356d0ec73635 BootID:77515c10-0352-4c6b-bba7-83e5133152ad Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:f2:41:ff Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:f2:41:ff Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:40:3c:35 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:24:87:97 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:56:3d:c2 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:5f:78:47 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:9e:e9:34:a3:56:0f Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:62:60:d0:5a:0c:8d Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.918239 4923 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.918431 4923 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.920007 4923 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.920240 4923 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.920285 4923 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.920595 4923 topology_manager.go:138] "Creating topology manager with none policy" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.920610 4923 container_manager_linux.go:303] "Creating device plugin manager" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.921562 4923 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.921601 4923 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.921817 4923 state_mem.go:36] "Initialized new in-memory state store" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.921925 4923 server.go:1245] "Using root directory" path="/var/lib/kubelet" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.926429 4923 kubelet.go:418] "Attempting to sync node with API server" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.926457 4923 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.926492 4923 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.926509 4923 kubelet.go:324] "Adding apiserver pod source" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.926564 4923 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.930681 4923 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.931498 4923 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.933886 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.933908 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Mar 11 13:22:32 crc kubenswrapper[4923]: E0311 13:22:32.934019 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.111:6443: connect: connection refused" logger="UnhandledError" Mar 11 13:22:32 crc kubenswrapper[4923]: E0311 13:22:32.934027 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.111:6443: connect: connection refused" logger="UnhandledError" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.934112 4923 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.935844 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.935874 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.935883 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.935891 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.935905 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.935915 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.935924 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.935937 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.935948 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.935958 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.935969 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.935978 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.937981 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.938529 4923 server.go:1280] "Started kubelet" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.938992 4923 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.939057 4923 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.939815 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.940647 4923 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 11 13:22:32 crc systemd[1]: Started Kubernetes Kubelet. Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.945721 4923 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.945773 4923 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 11 13:22:32 crc kubenswrapper[4923]: E0311 13:22:32.946962 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.947002 4923 volume_manager.go:287] "The desired_state_of_world populator starts" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.947949 4923 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.948264 4923 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.948896 4923 factory.go:55] Registering systemd factory Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.948947 4923 factory.go:221] Registration of the systemd container factory successfully Mar 11 13:22:32 crc kubenswrapper[4923]: E0311 13:22:32.951276 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.111:6443: connect: connection refused" interval="200ms" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.956665 4923 factory.go:153] Registering CRI-O factory Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.956740 4923 factory.go:221] Registration of the crio container factory successfully Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.956937 4923 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.956983 4923 factory.go:103] Registering Raw factory Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.957008 4923 manager.go:1196] Started watching for new ooms in manager Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.957546 4923 server.go:460] "Adding debug handlers to kubelet server" Mar 11 13:22:32 crc kubenswrapper[4923]: W0311 13:22:32.957582 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Mar 11 13:22:32 crc kubenswrapper[4923]: E0311 13:22:32.957731 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.111:6443: connect: connection refused" logger="UnhandledError" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.957826 4923 manager.go:319] Starting recovery of all containers Mar 11 13:22:32 crc kubenswrapper[4923]: E0311 13:22:32.958946 4923 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.111:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189bcc2893102e53 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:32.938491475 +0000 UTC m=+0.620393189,LastTimestamp:2026-03-11 13:22:32.938491475 +0000 UTC m=+0.620393189,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.971433 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.971537 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.971572 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.971600 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.971625 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.971650 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.971673 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.971700 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.971729 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.971754 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.971778 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.971801 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.971825 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.971854 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.971873 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.971895 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.971913 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.971933 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.971951 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.971971 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.971991 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972013 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972034 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972096 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972119 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972178 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972259 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972290 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972312 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972335 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972395 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972425 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972451 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972476 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972501 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972527 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972552 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972575 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972605 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972629 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972655 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972679 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972703 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972726 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972752 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972777 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972803 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972828 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972854 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972878 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972903 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972926 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972959 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.972985 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.973011 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.973037 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.979430 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.980419 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.980554 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.980652 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.980685 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.980714 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.980737 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.980763 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.980784 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.980811 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.981009 4923 manager.go:324] Recovery completed Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.981007 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.981363 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.981525 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.981682 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.981846 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.981979 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.982104 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.982210 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.982338 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.982499 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.982726 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.982900 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.983006 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.983115 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.983221 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.983337 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.983587 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.983714 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.983844 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.985102 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.985220 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.985331 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.985478 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.985594 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.985706 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.985820 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.985942 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.986056 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.986176 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.986675 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.986913 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.987021 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.987125 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.987221 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.987313 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.987460 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.987562 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.987677 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.987788 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.987894 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.988017 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.988131 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.988237 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.988386 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.988483 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.988564 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.988675 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.988764 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.988855 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.988934 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.989012 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.989099 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.989175 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.989259 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.992721 4923 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.992843 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.992930 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.993008 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.993103 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.993212 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.993312 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.993437 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.993527 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.993615 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.993706 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.993828 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.993935 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.994029 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.994112 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.994209 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.994292 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.994435 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.994529 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.994607 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.994699 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.994780 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.994858 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.994937 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.995013 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.995096 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.995175 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.995252 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.995327 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.995450 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.995532 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.995626 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.995744 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.995844 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.995948 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996029 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996117 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.993993 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996250 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996558 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996635 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996654 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996672 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996688 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996705 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996726 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996743 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996766 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996784 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996799 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996853 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996870 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996883 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996908 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996923 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996937 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996954 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996969 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.996985 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997001 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997015 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997031 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997046 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997064 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997078 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997092 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997108 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997124 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997141 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997161 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997176 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997193 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997209 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997229 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997243 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997262 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997279 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997295 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997310 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997325 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997390 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997409 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997426 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997442 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997462 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997483 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997504 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997523 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997542 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997559 4923 reconstruct.go:97] "Volume reconstruction finished" Mar 11 13:22:32 crc kubenswrapper[4923]: I0311 13:22:32.997573 4923 reconciler.go:26] "Reconciler: start to sync state" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.000199 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.000335 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.000406 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.001383 4923 cpu_manager.go:225] "Starting CPU manager" policy="none" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.001408 4923 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.001586 4923 state_mem.go:36] "Initialized new in-memory state store" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.017093 4923 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.019729 4923 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.020515 4923 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.020610 4923 kubelet.go:2335] "Starting kubelet main sync loop" Mar 11 13:22:33 crc kubenswrapper[4923]: E0311 13:22:33.020859 4923 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.022312 4923 policy_none.go:49] "None policy: Start" Mar 11 13:22:33 crc kubenswrapper[4923]: W0311 13:22:33.022620 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Mar 11 13:22:33 crc kubenswrapper[4923]: E0311 13:22:33.022809 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.111:6443: connect: connection refused" logger="UnhandledError" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.023479 4923 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.023536 4923 state_mem.go:35] "Initializing new in-memory state store" Mar 11 13:22:33 crc kubenswrapper[4923]: E0311 13:22:33.047992 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.089001 4923 manager.go:334] "Starting Device Plugin manager" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.089069 4923 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.089088 4923 server.go:79] "Starting device plugin registration server" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.089734 4923 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.089763 4923 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.090493 4923 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.090716 4923 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.090734 4923 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 11 13:22:33 crc kubenswrapper[4923]: E0311 13:22:33.097088 4923 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.121858 4923 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.121982 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.124994 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.125046 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.125061 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.125247 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.125507 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.125560 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.126404 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.126431 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.126445 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.126508 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.126540 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.126552 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.126582 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.126797 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.126895 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.127313 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.127366 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.127380 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.127520 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.127645 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.127686 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.128317 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.128377 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.128393 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.128534 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.128642 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.128658 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.128682 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.128698 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.128719 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.129035 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.129065 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.129076 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.129257 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.129280 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.129292 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.129452 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.129483 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.130213 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.130227 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.130280 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.130306 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.130248 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.130363 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:33 crc kubenswrapper[4923]: E0311 13:22:33.151974 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.111:6443: connect: connection refused" interval="400ms" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.189940 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.191244 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.191282 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.191294 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.191319 4923 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 11 13:22:33 crc kubenswrapper[4923]: E0311 13:22:33.191915 4923 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.111:6443: connect: connection refused" node="crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.200102 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.200159 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.200225 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.200286 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.200316 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.200387 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.200433 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.200463 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.200518 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.200548 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.200580 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.200624 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.200675 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.200735 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.200774 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.302383 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.302461 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.302499 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.302539 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.302575 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.302601 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.302626 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.302623 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.302671 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.302741 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.302652 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.302745 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.302785 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.302706 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.302811 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.302844 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.302847 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.302887 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.302891 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.302952 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.303047 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.303105 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.303378 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.303246 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.303510 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.303610 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.303688 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.303882 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.303946 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.303975 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.392668 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.394776 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.394847 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.394873 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.394912 4923 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 11 13:22:33 crc kubenswrapper[4923]: E0311 13:22:33.395522 4923 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.111:6443: connect: connection refused" node="crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.469963 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.477652 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.507111 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: W0311 13:22:33.531554 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-b93d40c4a0b72ce65f665c81aaa968432aa0b81041217e03544d27109058f9b5 WatchSource:0}: Error finding container b93d40c4a0b72ce65f665c81aaa968432aa0b81041217e03544d27109058f9b5: Status 404 returned error can't find the container with id b93d40c4a0b72ce65f665c81aaa968432aa0b81041217e03544d27109058f9b5 Mar 11 13:22:33 crc kubenswrapper[4923]: W0311 13:22:33.533166 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-a230b562b36842c52e6edec3d155e27eacb66f6156c6ff31833ab943a4f29eb9 WatchSource:0}: Error finding container a230b562b36842c52e6edec3d155e27eacb66f6156c6ff31833ab943a4f29eb9: Status 404 returned error can't find the container with id a230b562b36842c52e6edec3d155e27eacb66f6156c6ff31833ab943a4f29eb9 Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.535658 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: W0311 13:22:33.542303 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-3aafe25243dbe583980a4c5e7fe5d6dc0e05649eeabff5088ec0ace0fbd2e399 WatchSource:0}: Error finding container 3aafe25243dbe583980a4c5e7fe5d6dc0e05649eeabff5088ec0ace0fbd2e399: Status 404 returned error can't find the container with id 3aafe25243dbe583980a4c5e7fe5d6dc0e05649eeabff5088ec0ace0fbd2e399 Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.543196 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 11 13:22:33 crc kubenswrapper[4923]: E0311 13:22:33.553726 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.111:6443: connect: connection refused" interval="800ms" Mar 11 13:22:33 crc kubenswrapper[4923]: W0311 13:22:33.559868 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-ebc4a8bc3f53e2bfc8f6cb52e4cd4bf912b10fac3185792500153e76e33e3519 WatchSource:0}: Error finding container ebc4a8bc3f53e2bfc8f6cb52e4cd4bf912b10fac3185792500153e76e33e3519: Status 404 returned error can't find the container with id ebc4a8bc3f53e2bfc8f6cb52e4cd4bf912b10fac3185792500153e76e33e3519 Mar 11 13:22:33 crc kubenswrapper[4923]: W0311 13:22:33.570036 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-744e8b2623489ce205b414959fd544e3bed5173789c9df199e5a042653180534 WatchSource:0}: Error finding container 744e8b2623489ce205b414959fd544e3bed5173789c9df199e5a042653180534: Status 404 returned error can't find the container with id 744e8b2623489ce205b414959fd544e3bed5173789c9df199e5a042653180534 Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.796252 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.797787 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.797821 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.797831 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.797855 4923 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 11 13:22:33 crc kubenswrapper[4923]: E0311 13:22:33.798413 4923 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.111:6443: connect: connection refused" node="crc" Mar 11 13:22:33 crc kubenswrapper[4923]: I0311 13:22:33.941091 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Mar 11 13:22:34 crc kubenswrapper[4923]: I0311 13:22:34.026233 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"a230b562b36842c52e6edec3d155e27eacb66f6156c6ff31833ab943a4f29eb9"} Mar 11 13:22:34 crc kubenswrapper[4923]: I0311 13:22:34.027463 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"744e8b2623489ce205b414959fd544e3bed5173789c9df199e5a042653180534"} Mar 11 13:22:34 crc kubenswrapper[4923]: I0311 13:22:34.029072 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ebc4a8bc3f53e2bfc8f6cb52e4cd4bf912b10fac3185792500153e76e33e3519"} Mar 11 13:22:34 crc kubenswrapper[4923]: I0311 13:22:34.032723 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3aafe25243dbe583980a4c5e7fe5d6dc0e05649eeabff5088ec0ace0fbd2e399"} Mar 11 13:22:34 crc kubenswrapper[4923]: I0311 13:22:34.035996 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b93d40c4a0b72ce65f665c81aaa968432aa0b81041217e03544d27109058f9b5"} Mar 11 13:22:34 crc kubenswrapper[4923]: W0311 13:22:34.084142 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Mar 11 13:22:34 crc kubenswrapper[4923]: E0311 13:22:34.084235 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.111:6443: connect: connection refused" logger="UnhandledError" Mar 11 13:22:34 crc kubenswrapper[4923]: W0311 13:22:34.297188 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Mar 11 13:22:34 crc kubenswrapper[4923]: E0311 13:22:34.297620 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.111:6443: connect: connection refused" logger="UnhandledError" Mar 11 13:22:34 crc kubenswrapper[4923]: E0311 13:22:34.354735 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.111:6443: connect: connection refused" interval="1.6s" Mar 11 13:22:34 crc kubenswrapper[4923]: W0311 13:22:34.434101 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Mar 11 13:22:34 crc kubenswrapper[4923]: E0311 13:22:34.434273 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.111:6443: connect: connection refused" logger="UnhandledError" Mar 11 13:22:34 crc kubenswrapper[4923]: W0311 13:22:34.455322 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Mar 11 13:22:34 crc kubenswrapper[4923]: E0311 13:22:34.455443 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.111:6443: connect: connection refused" logger="UnhandledError" Mar 11 13:22:34 crc kubenswrapper[4923]: I0311 13:22:34.599417 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:34 crc kubenswrapper[4923]: I0311 13:22:34.601392 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:34 crc kubenswrapper[4923]: I0311 13:22:34.601448 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:34 crc kubenswrapper[4923]: I0311 13:22:34.601467 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:34 crc kubenswrapper[4923]: I0311 13:22:34.601505 4923 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 11 13:22:34 crc kubenswrapper[4923]: E0311 13:22:34.602138 4923 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.111:6443: connect: connection refused" node="crc" Mar 11 13:22:34 crc kubenswrapper[4923]: I0311 13:22:34.940953 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.009164 4923 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 11 13:22:35 crc kubenswrapper[4923]: E0311 13:22:35.010862 4923 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.111:6443: connect: connection refused" logger="UnhandledError" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.044106 4923 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="8713d48e2f4aad275760a5dac52cb2f7535da36c30c432162ba0c4e07fe726fd" exitCode=0 Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.044228 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"8713d48e2f4aad275760a5dac52cb2f7535da36c30c432162ba0c4e07fe726fd"} Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.044233 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.045817 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.045884 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.045897 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.046812 4923 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="d7ab99b63ef6d128d1e0e1a4b212710607d61e3d9a6de47bbc9fb148b0057509" exitCode=0 Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.046947 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"d7ab99b63ef6d128d1e0e1a4b212710607d61e3d9a6de47bbc9fb148b0057509"} Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.046972 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.048093 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.048127 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.048140 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.051078 4923 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7a46ea6935fa799f77d982717d4cbf69da04eee2f374af235d5876e79f32cdad" exitCode=0 Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.051167 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7a46ea6935fa799f77d982717d4cbf69da04eee2f374af235d5876e79f32cdad"} Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.051258 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.053029 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.053068 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.053087 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.056070 4923 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6e1f00d5974fffbe97d7d9d7dc2c8eab8a73b272ee66d9633c90ca97bd23ded1" exitCode=0 Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.056217 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"6e1f00d5974fffbe97d7d9d7dc2c8eab8a73b272ee66d9633c90ca97bd23ded1"} Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.056302 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.057713 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.057769 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.057789 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.061300 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"841b100c0220f381d9321eb3fe1cd4d885f225dff9556c6e3e888c393c875093"} Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.061404 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3c858a12f95f92b452180d16a20b6b3e15cbc591936efc4232bf45e4c876d3d0"} Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.061432 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"9c8bde01ba2b353875bbe2fad09608080e6ceeebbc890899a7b690a85de76acf"} Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.061453 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4b29200103f0d31fe398058b057f9843b9eec00edc4d3ed2da9f794618ca50c8"} Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.061455 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.062518 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.062932 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.062989 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.063013 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.063499 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.063550 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.063570 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:35 crc kubenswrapper[4923]: I0311 13:22:35.940650 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Mar 11 13:22:35 crc kubenswrapper[4923]: E0311 13:22:35.955521 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.111:6443: connect: connection refused" interval="3.2s" Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.068747 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"9872ea2a59b6fa0eb653661f2c6c9468cf52595e2b903489f288fd896a3eace1"} Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.068811 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.069989 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.070042 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.070062 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.072101 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a5d8c9ca3130b14594ab74918053c30b06ac6898980eb73e45a461d54c62b816"} Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.072139 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.072166 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"720e08c3e44856821836d7b55a14b93120eb5ceaeeda3a409b10661926394b55"} Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.072185 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"89454c2d00ab987f6fea87c7bb8b6af7f5088eba22075cf4f35956afab1a3e50"} Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.073807 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.073852 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.073868 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.075750 4923 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="58714b4fa3e5f9b226b4132d3340f3730ea44224f88f8377d566e2f5fb311e1a" exitCode=0 Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.075797 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"58714b4fa3e5f9b226b4132d3340f3730ea44224f88f8377d566e2f5fb311e1a"} Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.075977 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.076972 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.077014 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.077029 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.079987 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8310025c89cd82e434fa820785f24adeb870dc4c436e02258252b3a8493dc6ca"} Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.080036 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2a2537d234646cc42128c4e97220f789c9bcbe25c060c0f488aa820e3f611848"} Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.080051 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b1e1d3884306e5cb7b7d8e416ba17caa145f530c0f6d7b35a81ef10f4a79fadb"} Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.080063 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.080061 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b865dff977394535b945fbacba9ca628914f266d5e6a4bf1281a6771d884f622"} Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.082554 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.082604 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.082620 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.203214 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.204648 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.204722 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.204747 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:36 crc kubenswrapper[4923]: I0311 13:22:36.204792 4923 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 11 13:22:36 crc kubenswrapper[4923]: E0311 13:22:36.208368 4923 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.111:6443: connect: connection refused" node="crc" Mar 11 13:22:36 crc kubenswrapper[4923]: W0311 13:22:36.556760 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.111:6443: connect: connection refused Mar 11 13:22:36 crc kubenswrapper[4923]: E0311 13:22:36.556823 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.111:6443: connect: connection refused" logger="UnhandledError" Mar 11 13:22:37 crc kubenswrapper[4923]: I0311 13:22:37.039702 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 11 13:22:37 crc kubenswrapper[4923]: I0311 13:22:37.086030 4923 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1ce9eacb96d43d95757bed37da8ab34139a576c89086763bdc7c418d5317f3d2" exitCode=0 Mar 11 13:22:37 crc kubenswrapper[4923]: I0311 13:22:37.086141 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1ce9eacb96d43d95757bed37da8ab34139a576c89086763bdc7c418d5317f3d2"} Mar 11 13:22:37 crc kubenswrapper[4923]: I0311 13:22:37.086248 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:37 crc kubenswrapper[4923]: I0311 13:22:37.087717 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:37 crc kubenswrapper[4923]: I0311 13:22:37.087767 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:37 crc kubenswrapper[4923]: I0311 13:22:37.087785 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:37 crc kubenswrapper[4923]: I0311 13:22:37.092486 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6e4bc79805d539424fb0a4df3455f89c90f24374c38fd977b1f65346902990e0"} Mar 11 13:22:37 crc kubenswrapper[4923]: I0311 13:22:37.092511 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:37 crc kubenswrapper[4923]: I0311 13:22:37.092580 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:37 crc kubenswrapper[4923]: I0311 13:22:37.092615 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:37 crc kubenswrapper[4923]: I0311 13:22:37.093964 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:37 crc kubenswrapper[4923]: I0311 13:22:37.094018 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:37 crc kubenswrapper[4923]: I0311 13:22:37.094038 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:37 crc kubenswrapper[4923]: I0311 13:22:37.094265 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:37 crc kubenswrapper[4923]: I0311 13:22:37.094302 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:37 crc kubenswrapper[4923]: I0311 13:22:37.094322 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:37 crc kubenswrapper[4923]: I0311 13:22:37.095031 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:37 crc kubenswrapper[4923]: I0311 13:22:37.095077 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:37 crc kubenswrapper[4923]: I0311 13:22:37.095094 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:37 crc kubenswrapper[4923]: I0311 13:22:37.522321 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:22:38 crc kubenswrapper[4923]: I0311 13:22:38.100508 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a22ef3c797d7ad147bd7f034e440f21f5f87a605a81b93c7f58e6186afc8ffc4"} Mar 11 13:22:38 crc kubenswrapper[4923]: I0311 13:22:38.100568 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5da976d58b9b94e4a7f6eda9f828b2373dd0ece32e49f2610208081a50dd5be9"} Mar 11 13:22:38 crc kubenswrapper[4923]: I0311 13:22:38.100601 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:22:38 crc kubenswrapper[4923]: I0311 13:22:38.100623 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"83495046b69d717a5d75dde645413b379383729bc32cdc3f7a9914c75e93d3fe"} Mar 11 13:22:38 crc kubenswrapper[4923]: I0311 13:22:38.100577 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:38 crc kubenswrapper[4923]: I0311 13:22:38.100672 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:38 crc kubenswrapper[4923]: I0311 13:22:38.102197 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:38 crc kubenswrapper[4923]: I0311 13:22:38.102246 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:38 crc kubenswrapper[4923]: I0311 13:22:38.102263 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:38 crc kubenswrapper[4923]: I0311 13:22:38.102286 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:38 crc kubenswrapper[4923]: I0311 13:22:38.102361 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:38 crc kubenswrapper[4923]: I0311 13:22:38.102380 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:39 crc kubenswrapper[4923]: I0311 13:22:39.109395 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7a4389eb4cab063f10a4aa2305f758287ab414d18262d4def48a1620c676f33f"} Mar 11 13:22:39 crc kubenswrapper[4923]: I0311 13:22:39.109476 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8138e5b2ed09f30853f1ea2bd908ab9a108b9015f09f163ca4ed488c1079bf89"} Mar 11 13:22:39 crc kubenswrapper[4923]: I0311 13:22:39.109485 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:39 crc kubenswrapper[4923]: I0311 13:22:39.109678 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:39 crc kubenswrapper[4923]: I0311 13:22:39.110873 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:39 crc kubenswrapper[4923]: I0311 13:22:39.110913 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:39 crc kubenswrapper[4923]: I0311 13:22:39.110925 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:39 crc kubenswrapper[4923]: I0311 13:22:39.110935 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:39 crc kubenswrapper[4923]: I0311 13:22:39.110971 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:39 crc kubenswrapper[4923]: I0311 13:22:39.110989 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:39 crc kubenswrapper[4923]: I0311 13:22:39.409287 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:39 crc kubenswrapper[4923]: I0311 13:22:39.410324 4923 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 11 13:22:39 crc kubenswrapper[4923]: I0311 13:22:39.411605 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:39 crc kubenswrapper[4923]: I0311 13:22:39.411683 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:39 crc kubenswrapper[4923]: I0311 13:22:39.411711 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:39 crc kubenswrapper[4923]: I0311 13:22:39.411746 4923 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 11 13:22:40 crc kubenswrapper[4923]: I0311 13:22:40.035961 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:22:40 crc kubenswrapper[4923]: I0311 13:22:40.036225 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:40 crc kubenswrapper[4923]: I0311 13:22:40.038036 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:40 crc kubenswrapper[4923]: I0311 13:22:40.038102 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:40 crc kubenswrapper[4923]: I0311 13:22:40.038125 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:40 crc kubenswrapper[4923]: I0311 13:22:40.112879 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:40 crc kubenswrapper[4923]: I0311 13:22:40.114720 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:40 crc kubenswrapper[4923]: I0311 13:22:40.114774 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:40 crc kubenswrapper[4923]: I0311 13:22:40.114793 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:40 crc kubenswrapper[4923]: I0311 13:22:40.712264 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:22:40 crc kubenswrapper[4923]: I0311 13:22:40.712559 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:40 crc kubenswrapper[4923]: I0311 13:22:40.714404 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:40 crc kubenswrapper[4923]: I0311 13:22:40.714460 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:40 crc kubenswrapper[4923]: I0311 13:22:40.714485 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:41 crc kubenswrapper[4923]: I0311 13:22:41.669489 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:22:41 crc kubenswrapper[4923]: I0311 13:22:41.669706 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:41 crc kubenswrapper[4923]: I0311 13:22:41.671435 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:41 crc kubenswrapper[4923]: I0311 13:22:41.671501 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:41 crc kubenswrapper[4923]: I0311 13:22:41.671520 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:41 crc kubenswrapper[4923]: I0311 13:22:41.683142 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:22:42 crc kubenswrapper[4923]: I0311 13:22:42.118854 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:42 crc kubenswrapper[4923]: I0311 13:22:42.120128 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:42 crc kubenswrapper[4923]: I0311 13:22:42.120207 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:42 crc kubenswrapper[4923]: I0311 13:22:42.120226 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:43 crc kubenswrapper[4923]: E0311 13:22:43.097244 4923 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 11 13:22:43 crc kubenswrapper[4923]: I0311 13:22:43.666677 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 11 13:22:43 crc kubenswrapper[4923]: I0311 13:22:43.666844 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:43 crc kubenswrapper[4923]: I0311 13:22:43.668003 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:43 crc kubenswrapper[4923]: I0311 13:22:43.668037 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:43 crc kubenswrapper[4923]: I0311 13:22:43.668049 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:44 crc kubenswrapper[4923]: I0311 13:22:44.022481 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Mar 11 13:22:44 crc kubenswrapper[4923]: I0311 13:22:44.124420 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:44 crc kubenswrapper[4923]: I0311 13:22:44.125936 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:44 crc kubenswrapper[4923]: I0311 13:22:44.125990 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:44 crc kubenswrapper[4923]: I0311 13:22:44.126002 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:44 crc kubenswrapper[4923]: I0311 13:22:44.253009 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:22:44 crc kubenswrapper[4923]: I0311 13:22:44.253185 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:44 crc kubenswrapper[4923]: I0311 13:22:44.254912 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:44 crc kubenswrapper[4923]: I0311 13:22:44.254965 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:44 crc kubenswrapper[4923]: I0311 13:22:44.254977 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:44 crc kubenswrapper[4923]: I0311 13:22:44.260843 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:22:44 crc kubenswrapper[4923]: I0311 13:22:44.669498 4923 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 11 13:22:44 crc kubenswrapper[4923]: I0311 13:22:44.669575 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 11 13:22:45 crc kubenswrapper[4923]: I0311 13:22:45.127988 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:45 crc kubenswrapper[4923]: I0311 13:22:45.130071 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:45 crc kubenswrapper[4923]: I0311 13:22:45.130135 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:45 crc kubenswrapper[4923]: I0311 13:22:45.130182 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:45 crc kubenswrapper[4923]: I0311 13:22:45.138795 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:22:46 crc kubenswrapper[4923]: I0311 13:22:46.131102 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:46 crc kubenswrapper[4923]: I0311 13:22:46.132747 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:46 crc kubenswrapper[4923]: I0311 13:22:46.132805 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:46 crc kubenswrapper[4923]: I0311 13:22:46.132823 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:46 crc kubenswrapper[4923]: W0311 13:22:46.802255 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 11 13:22:46 crc kubenswrapper[4923]: I0311 13:22:46.802375 4923 trace.go:236] Trace[2030441838]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Mar-2026 13:22:36.801) (total time: 10000ms): Mar 11 13:22:46 crc kubenswrapper[4923]: Trace[2030441838]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10000ms (13:22:46.802) Mar 11 13:22:46 crc kubenswrapper[4923]: Trace[2030441838]: [10.000872869s] [10.000872869s] END Mar 11 13:22:46 crc kubenswrapper[4923]: E0311 13:22:46.802402 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 11 13:22:46 crc kubenswrapper[4923]: I0311 13:22:46.940828 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Mar 11 13:22:46 crc kubenswrapper[4923]: W0311 13:22:46.949186 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 11 13:22:46 crc kubenswrapper[4923]: I0311 13:22:46.949299 4923 trace.go:236] Trace[695633923]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Mar-2026 13:22:36.947) (total time: 10001ms): Mar 11 13:22:46 crc kubenswrapper[4923]: Trace[695633923]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (13:22:46.949) Mar 11 13:22:46 crc kubenswrapper[4923]: Trace[695633923]: [10.001583855s] [10.001583855s] END Mar 11 13:22:46 crc kubenswrapper[4923]: E0311 13:22:46.949325 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 11 13:22:47 crc kubenswrapper[4923]: W0311 13:22:47.013973 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 11 13:22:47 crc kubenswrapper[4923]: I0311 13:22:47.014078 4923 trace.go:236] Trace[1586740013]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Mar-2026 13:22:37.012) (total time: 10001ms): Mar 11 13:22:47 crc kubenswrapper[4923]: Trace[1586740013]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (13:22:47.013) Mar 11 13:22:47 crc kubenswrapper[4923]: Trace[1586740013]: [10.001569377s] [10.001569377s] END Mar 11 13:22:47 crc kubenswrapper[4923]: E0311 13:22:47.014114 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 11 13:22:47 crc kubenswrapper[4923]: E0311 13:22:47.337920 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:47Z is after 2026-02-23T05:33:13Z" interval="6.4s" Mar 11 13:22:47 crc kubenswrapper[4923]: W0311 13:22:47.343203 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:47Z is after 2026-02-23T05:33:13Z Mar 11 13:22:47 crc kubenswrapper[4923]: E0311 13:22:47.343302 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:47Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 11 13:22:47 crc kubenswrapper[4923]: E0311 13:22:47.343539 4923 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:47Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 11 13:22:47 crc kubenswrapper[4923]: E0311 13:22:47.343555 4923 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:47Z is after 2026-02-23T05:33:13Z" node="crc" Mar 11 13:22:47 crc kubenswrapper[4923]: I0311 13:22:47.346787 4923 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 11 13:22:47 crc kubenswrapper[4923]: I0311 13:22:47.346843 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 11 13:22:47 crc kubenswrapper[4923]: E0311 13:22:47.347563 4923 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:47Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189bcc2893102e53 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:32.938491475 +0000 UTC m=+0.620393189,LastTimestamp:2026-03-11 13:22:32.938491475 +0000 UTC m=+0.620393189,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:22:47 crc kubenswrapper[4923]: I0311 13:22:47.352181 4923 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 11 13:22:47 crc kubenswrapper[4923]: I0311 13:22:47.352217 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 11 13:22:47 crc kubenswrapper[4923]: I0311 13:22:47.556189 4923 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Mar 11 13:22:47 crc kubenswrapper[4923]: I0311 13:22:47.556295 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Mar 11 13:22:47 crc kubenswrapper[4923]: I0311 13:22:47.947833 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:47Z is after 2026-02-23T05:33:13Z Mar 11 13:22:48 crc kubenswrapper[4923]: I0311 13:22:48.138264 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 11 13:22:48 crc kubenswrapper[4923]: I0311 13:22:48.140378 4923 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6e4bc79805d539424fb0a4df3455f89c90f24374c38fd977b1f65346902990e0" exitCode=255 Mar 11 13:22:48 crc kubenswrapper[4923]: I0311 13:22:48.140417 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"6e4bc79805d539424fb0a4df3455f89c90f24374c38fd977b1f65346902990e0"} Mar 11 13:22:48 crc kubenswrapper[4923]: I0311 13:22:48.140559 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:48 crc kubenswrapper[4923]: I0311 13:22:48.141649 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:48 crc kubenswrapper[4923]: I0311 13:22:48.141729 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:48 crc kubenswrapper[4923]: I0311 13:22:48.141756 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:48 crc kubenswrapper[4923]: I0311 13:22:48.142644 4923 scope.go:117] "RemoveContainer" containerID="6e4bc79805d539424fb0a4df3455f89c90f24374c38fd977b1f65346902990e0" Mar 11 13:22:48 crc kubenswrapper[4923]: I0311 13:22:48.948128 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:48Z is after 2026-02-23T05:33:13Z Mar 11 13:22:49 crc kubenswrapper[4923]: I0311 13:22:49.145806 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 11 13:22:49 crc kubenswrapper[4923]: I0311 13:22:49.147816 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0de3d701084f9234db9bb4b76d24ce0d3d40c90cd3422ed49b522207d498de90"} Mar 11 13:22:49 crc kubenswrapper[4923]: I0311 13:22:49.148058 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:49 crc kubenswrapper[4923]: I0311 13:22:49.149486 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:49 crc kubenswrapper[4923]: I0311 13:22:49.149539 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:49 crc kubenswrapper[4923]: I0311 13:22:49.149559 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:49 crc kubenswrapper[4923]: I0311 13:22:49.946738 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:49Z is after 2026-02-23T05:33:13Z Mar 11 13:22:50 crc kubenswrapper[4923]: I0311 13:22:50.153822 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 11 13:22:50 crc kubenswrapper[4923]: I0311 13:22:50.154600 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 11 13:22:50 crc kubenswrapper[4923]: I0311 13:22:50.157894 4923 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0de3d701084f9234db9bb4b76d24ce0d3d40c90cd3422ed49b522207d498de90" exitCode=255 Mar 11 13:22:50 crc kubenswrapper[4923]: I0311 13:22:50.157971 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"0de3d701084f9234db9bb4b76d24ce0d3d40c90cd3422ed49b522207d498de90"} Mar 11 13:22:50 crc kubenswrapper[4923]: I0311 13:22:50.158063 4923 scope.go:117] "RemoveContainer" containerID="6e4bc79805d539424fb0a4df3455f89c90f24374c38fd977b1f65346902990e0" Mar 11 13:22:50 crc kubenswrapper[4923]: I0311 13:22:50.158264 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:50 crc kubenswrapper[4923]: I0311 13:22:50.160034 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:50 crc kubenswrapper[4923]: I0311 13:22:50.160120 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:50 crc kubenswrapper[4923]: I0311 13:22:50.160141 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:50 crc kubenswrapper[4923]: I0311 13:22:50.162382 4923 scope.go:117] "RemoveContainer" containerID="0de3d701084f9234db9bb4b76d24ce0d3d40c90cd3422ed49b522207d498de90" Mar 11 13:22:50 crc kubenswrapper[4923]: E0311 13:22:50.162769 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 11 13:22:50 crc kubenswrapper[4923]: I0311 13:22:50.722111 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:22:50 crc kubenswrapper[4923]: I0311 13:22:50.945337 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:50Z is after 2026-02-23T05:33:13Z Mar 11 13:22:51 crc kubenswrapper[4923]: I0311 13:22:51.163846 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 11 13:22:51 crc kubenswrapper[4923]: I0311 13:22:51.167952 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:51 crc kubenswrapper[4923]: I0311 13:22:51.169471 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:51 crc kubenswrapper[4923]: I0311 13:22:51.169699 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:51 crc kubenswrapper[4923]: I0311 13:22:51.169894 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:51 crc kubenswrapper[4923]: I0311 13:22:51.170852 4923 scope.go:117] "RemoveContainer" containerID="0de3d701084f9234db9bb4b76d24ce0d3d40c90cd3422ed49b522207d498de90" Mar 11 13:22:51 crc kubenswrapper[4923]: E0311 13:22:51.171317 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 11 13:22:51 crc kubenswrapper[4923]: I0311 13:22:51.174738 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:22:51 crc kubenswrapper[4923]: W0311 13:22:51.242608 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:51Z is after 2026-02-23T05:33:13Z Mar 11 13:22:51 crc kubenswrapper[4923]: E0311 13:22:51.242705 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:51Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 11 13:22:51 crc kubenswrapper[4923]: W0311 13:22:51.693898 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:51Z is after 2026-02-23T05:33:13Z Mar 11 13:22:51 crc kubenswrapper[4923]: E0311 13:22:51.694005 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:51Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 11 13:22:51 crc kubenswrapper[4923]: I0311 13:22:51.945187 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:51Z is after 2026-02-23T05:33:13Z Mar 11 13:22:52 crc kubenswrapper[4923]: I0311 13:22:52.170677 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:52 crc kubenswrapper[4923]: I0311 13:22:52.172032 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:52 crc kubenswrapper[4923]: I0311 13:22:52.172090 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:52 crc kubenswrapper[4923]: I0311 13:22:52.172112 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:52 crc kubenswrapper[4923]: I0311 13:22:52.173131 4923 scope.go:117] "RemoveContainer" containerID="0de3d701084f9234db9bb4b76d24ce0d3d40c90cd3422ed49b522207d498de90" Mar 11 13:22:52 crc kubenswrapper[4923]: E0311 13:22:52.173703 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 11 13:22:52 crc kubenswrapper[4923]: W0311 13:22:52.829740 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:52Z is after 2026-02-23T05:33:13Z Mar 11 13:22:52 crc kubenswrapper[4923]: E0311 13:22:52.829844 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:52Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 11 13:22:52 crc kubenswrapper[4923]: I0311 13:22:52.946476 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:52Z is after 2026-02-23T05:33:13Z Mar 11 13:22:53 crc kubenswrapper[4923]: E0311 13:22:53.097423 4923 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 11 13:22:53 crc kubenswrapper[4923]: I0311 13:22:53.696370 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:22:53 crc kubenswrapper[4923]: I0311 13:22:53.696940 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:53 crc kubenswrapper[4923]: I0311 13:22:53.698979 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:53 crc kubenswrapper[4923]: I0311 13:22:53.699077 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:53 crc kubenswrapper[4923]: I0311 13:22:53.699101 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:53 crc kubenswrapper[4923]: I0311 13:22:53.700255 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 11 13:22:53 crc kubenswrapper[4923]: I0311 13:22:53.700410 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:53 crc kubenswrapper[4923]: I0311 13:22:53.700443 4923 scope.go:117] "RemoveContainer" containerID="0de3d701084f9234db9bb4b76d24ce0d3d40c90cd3422ed49b522207d498de90" Mar 11 13:22:53 crc kubenswrapper[4923]: E0311 13:22:53.700830 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 11 13:22:53 crc kubenswrapper[4923]: I0311 13:22:53.701930 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:53 crc kubenswrapper[4923]: I0311 13:22:53.701974 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:53 crc kubenswrapper[4923]: I0311 13:22:53.701988 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:53 crc kubenswrapper[4923]: I0311 13:22:53.720261 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 11 13:22:53 crc kubenswrapper[4923]: I0311 13:22:53.744744 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:53 crc kubenswrapper[4923]: E0311 13:22:53.744878 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:53Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 11 13:22:53 crc kubenswrapper[4923]: I0311 13:22:53.747416 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:53 crc kubenswrapper[4923]: I0311 13:22:53.747505 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:53 crc kubenswrapper[4923]: I0311 13:22:53.747525 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:53 crc kubenswrapper[4923]: I0311 13:22:53.747588 4923 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 11 13:22:53 crc kubenswrapper[4923]: E0311 13:22:53.753004 4923 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:53Z is after 2026-02-23T05:33:13Z" node="crc" Mar 11 13:22:53 crc kubenswrapper[4923]: I0311 13:22:53.945388 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:53Z is after 2026-02-23T05:33:13Z Mar 11 13:22:54 crc kubenswrapper[4923]: I0311 13:22:54.177306 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:54 crc kubenswrapper[4923]: I0311 13:22:54.178798 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:54 crc kubenswrapper[4923]: I0311 13:22:54.178853 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:54 crc kubenswrapper[4923]: I0311 13:22:54.178871 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:54 crc kubenswrapper[4923]: I0311 13:22:54.670107 4923 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 11 13:22:54 crc kubenswrapper[4923]: I0311 13:22:54.670203 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 11 13:22:54 crc kubenswrapper[4923]: I0311 13:22:54.946003 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:54Z is after 2026-02-23T05:33:13Z Mar 11 13:22:55 crc kubenswrapper[4923]: I0311 13:22:55.797891 4923 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 11 13:22:55 crc kubenswrapper[4923]: E0311 13:22:55.804179 4923 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:55Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 11 13:22:55 crc kubenswrapper[4923]: I0311 13:22:55.944859 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:55Z is after 2026-02-23T05:33:13Z Mar 11 13:22:56 crc kubenswrapper[4923]: I0311 13:22:56.945735 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:56Z is after 2026-02-23T05:33:13Z Mar 11 13:22:57 crc kubenswrapper[4923]: E0311 13:22:57.353537 4923 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:57Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189bcc2893102e53 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:32.938491475 +0000 UTC m=+0.620393189,LastTimestamp:2026-03-11 13:22:32.938491475 +0000 UTC m=+0.620393189,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:22:57 crc kubenswrapper[4923]: I0311 13:22:57.555770 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:22:57 crc kubenswrapper[4923]: I0311 13:22:57.556443 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:22:57 crc kubenswrapper[4923]: I0311 13:22:57.558049 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:22:57 crc kubenswrapper[4923]: I0311 13:22:57.558101 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:22:57 crc kubenswrapper[4923]: I0311 13:22:57.558122 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:22:57 crc kubenswrapper[4923]: I0311 13:22:57.558861 4923 scope.go:117] "RemoveContainer" containerID="0de3d701084f9234db9bb4b76d24ce0d3d40c90cd3422ed49b522207d498de90" Mar 11 13:22:57 crc kubenswrapper[4923]: E0311 13:22:57.559127 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 11 13:22:57 crc kubenswrapper[4923]: W0311 13:22:57.924533 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:57Z is after 2026-02-23T05:33:13Z Mar 11 13:22:57 crc kubenswrapper[4923]: E0311 13:22:57.924645 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:57Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 11 13:22:57 crc kubenswrapper[4923]: I0311 13:22:57.945428 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:57Z is after 2026-02-23T05:33:13Z Mar 11 13:22:58 crc kubenswrapper[4923]: W0311 13:22:58.510862 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:58Z is after 2026-02-23T05:33:13Z Mar 11 13:22:58 crc kubenswrapper[4923]: E0311 13:22:58.510992 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:58Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 11 13:22:58 crc kubenswrapper[4923]: I0311 13:22:58.945972 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:58Z is after 2026-02-23T05:33:13Z Mar 11 13:22:59 crc kubenswrapper[4923]: W0311 13:22:59.158751 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:59Z is after 2026-02-23T05:33:13Z Mar 11 13:22:59 crc kubenswrapper[4923]: E0311 13:22:59.158868 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:59Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 11 13:22:59 crc kubenswrapper[4923]: I0311 13:22:59.949588 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:22:59Z is after 2026-02-23T05:33:13Z Mar 11 13:23:00 crc kubenswrapper[4923]: E0311 13:23:00.752400 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:23:00Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 11 13:23:00 crc kubenswrapper[4923]: I0311 13:23:00.753555 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:00 crc kubenswrapper[4923]: I0311 13:23:00.755863 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:00 crc kubenswrapper[4923]: I0311 13:23:00.755917 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:00 crc kubenswrapper[4923]: I0311 13:23:00.755935 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:00 crc kubenswrapper[4923]: I0311 13:23:00.755975 4923 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 11 13:23:00 crc kubenswrapper[4923]: E0311 13:23:00.761697 4923 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:23:00Z is after 2026-02-23T05:33:13Z" node="crc" Mar 11 13:23:00 crc kubenswrapper[4923]: I0311 13:23:00.945529 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:23:00Z is after 2026-02-23T05:33:13Z Mar 11 13:23:01 crc kubenswrapper[4923]: I0311 13:23:01.946228 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:23:01Z is after 2026-02-23T05:33:13Z Mar 11 13:23:02 crc kubenswrapper[4923]: I0311 13:23:02.944699 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:23:02Z is after 2026-02-23T05:33:13Z Mar 11 13:23:03 crc kubenswrapper[4923]: E0311 13:23:03.097561 4923 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 11 13:23:03 crc kubenswrapper[4923]: I0311 13:23:03.946057 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:23:03Z is after 2026-02-23T05:33:13Z Mar 11 13:23:04 crc kubenswrapper[4923]: I0311 13:23:04.670948 4923 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 11 13:23:04 crc kubenswrapper[4923]: I0311 13:23:04.671060 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 11 13:23:04 crc kubenswrapper[4923]: I0311 13:23:04.671148 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:23:04 crc kubenswrapper[4923]: I0311 13:23:04.671336 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:04 crc kubenswrapper[4923]: I0311 13:23:04.673961 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:04 crc kubenswrapper[4923]: I0311 13:23:04.674116 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:04 crc kubenswrapper[4923]: I0311 13:23:04.674151 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:04 crc kubenswrapper[4923]: I0311 13:23:04.676145 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"9c8bde01ba2b353875bbe2fad09608080e6ceeebbc890899a7b690a85de76acf"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 11 13:23:04 crc kubenswrapper[4923]: I0311 13:23:04.677231 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://9c8bde01ba2b353875bbe2fad09608080e6ceeebbc890899a7b690a85de76acf" gracePeriod=30 Mar 11 13:23:04 crc kubenswrapper[4923]: I0311 13:23:04.944043 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:23:04Z is after 2026-02-23T05:33:13Z Mar 11 13:23:04 crc kubenswrapper[4923]: W0311 13:23:04.948391 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:23:04Z is after 2026-02-23T05:33:13Z Mar 11 13:23:04 crc kubenswrapper[4923]: E0311 13:23:04.948494 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:23:04Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 11 13:23:05 crc kubenswrapper[4923]: I0311 13:23:05.212446 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 11 13:23:05 crc kubenswrapper[4923]: I0311 13:23:05.213131 4923 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="9c8bde01ba2b353875bbe2fad09608080e6ceeebbc890899a7b690a85de76acf" exitCode=255 Mar 11 13:23:05 crc kubenswrapper[4923]: I0311 13:23:05.213199 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"9c8bde01ba2b353875bbe2fad09608080e6ceeebbc890899a7b690a85de76acf"} Mar 11 13:23:05 crc kubenswrapper[4923]: I0311 13:23:05.213254 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0beb3682649fc24dbb6916120a56e14fb595d5898c39c2bfc65a93591c204a80"} Mar 11 13:23:05 crc kubenswrapper[4923]: I0311 13:23:05.213457 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:05 crc kubenswrapper[4923]: I0311 13:23:05.215301 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:05 crc kubenswrapper[4923]: I0311 13:23:05.215397 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:05 crc kubenswrapper[4923]: I0311 13:23:05.215425 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:05 crc kubenswrapper[4923]: I0311 13:23:05.946041 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:23:05Z is after 2026-02-23T05:33:13Z Mar 11 13:23:06 crc kubenswrapper[4923]: I0311 13:23:06.946152 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:23:06Z is after 2026-02-23T05:33:13Z Mar 11 13:23:07 crc kubenswrapper[4923]: E0311 13:23:07.359669 4923 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:23:07Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189bcc2893102e53 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:32.938491475 +0000 UTC m=+0.620393189,LastTimestamp:2026-03-11 13:22:32.938491475 +0000 UTC m=+0.620393189,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:07 crc kubenswrapper[4923]: E0311 13:23:07.757776 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:23:07Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 11 13:23:07 crc kubenswrapper[4923]: I0311 13:23:07.761824 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:07 crc kubenswrapper[4923]: I0311 13:23:07.763900 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:07 crc kubenswrapper[4923]: I0311 13:23:07.763957 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:07 crc kubenswrapper[4923]: I0311 13:23:07.763977 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:07 crc kubenswrapper[4923]: I0311 13:23:07.764020 4923 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 11 13:23:07 crc kubenswrapper[4923]: E0311 13:23:07.767699 4923 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:23:07Z is after 2026-02-23T05:33:13Z" node="crc" Mar 11 13:23:07 crc kubenswrapper[4923]: I0311 13:23:07.944136 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:23:07Z is after 2026-02-23T05:33:13Z Mar 11 13:23:08 crc kubenswrapper[4923]: I0311 13:23:08.945009 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:23:08Z is after 2026-02-23T05:33:13Z Mar 11 13:23:09 crc kubenswrapper[4923]: I0311 13:23:09.945328 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:23:09Z is after 2026-02-23T05:33:13Z Mar 11 13:23:10 crc kubenswrapper[4923]: I0311 13:23:10.036278 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:23:10 crc kubenswrapper[4923]: I0311 13:23:10.036648 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:10 crc kubenswrapper[4923]: I0311 13:23:10.039473 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:10 crc kubenswrapper[4923]: I0311 13:23:10.039569 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:10 crc kubenswrapper[4923]: I0311 13:23:10.039599 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:10 crc kubenswrapper[4923]: I0311 13:23:10.944200 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:23:10Z is after 2026-02-23T05:33:13Z Mar 11 13:23:11 crc kubenswrapper[4923]: I0311 13:23:11.670068 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:23:11 crc kubenswrapper[4923]: I0311 13:23:11.670653 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:11 crc kubenswrapper[4923]: I0311 13:23:11.673308 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:11 crc kubenswrapper[4923]: I0311 13:23:11.673465 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:11 crc kubenswrapper[4923]: I0311 13:23:11.673488 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:11 crc kubenswrapper[4923]: I0311 13:23:11.948393 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:12 crc kubenswrapper[4923]: I0311 13:23:12.012855 4923 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 11 13:23:12 crc kubenswrapper[4923]: I0311 13:23:12.022536 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:12 crc kubenswrapper[4923]: I0311 13:23:12.024577 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:12 crc kubenswrapper[4923]: I0311 13:23:12.024627 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:12 crc kubenswrapper[4923]: I0311 13:23:12.024645 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:12 crc kubenswrapper[4923]: I0311 13:23:12.025480 4923 scope.go:117] "RemoveContainer" containerID="0de3d701084f9234db9bb4b76d24ce0d3d40c90cd3422ed49b522207d498de90" Mar 11 13:23:12 crc kubenswrapper[4923]: I0311 13:23:12.035401 4923 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 11 13:23:12 crc kubenswrapper[4923]: I0311 13:23:12.948894 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:13 crc kubenswrapper[4923]: E0311 13:23:13.098669 4923 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 11 13:23:13 crc kubenswrapper[4923]: I0311 13:23:13.245387 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 11 13:23:13 crc kubenswrapper[4923]: I0311 13:23:13.247918 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3bf1a33b60de01496d3f642c6b988be5ce449a4c1a7b02919a4af52dab5d7b33"} Mar 11 13:23:13 crc kubenswrapper[4923]: I0311 13:23:13.248114 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:13 crc kubenswrapper[4923]: I0311 13:23:13.249585 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:13 crc kubenswrapper[4923]: I0311 13:23:13.249636 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:13 crc kubenswrapper[4923]: I0311 13:23:13.249653 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:13 crc kubenswrapper[4923]: I0311 13:23:13.696613 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:23:13 crc kubenswrapper[4923]: W0311 13:23:13.860502 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 11 13:23:13 crc kubenswrapper[4923]: E0311 13:23:13.860600 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 11 13:23:13 crc kubenswrapper[4923]: I0311 13:23:13.951811 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:14 crc kubenswrapper[4923]: W0311 13:23:14.183010 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:14 crc kubenswrapper[4923]: E0311 13:23:14.183086 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 11 13:23:14 crc kubenswrapper[4923]: I0311 13:23:14.253868 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 11 13:23:14 crc kubenswrapper[4923]: I0311 13:23:14.254559 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 11 13:23:14 crc kubenswrapper[4923]: I0311 13:23:14.257210 4923 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3bf1a33b60de01496d3f642c6b988be5ce449a4c1a7b02919a4af52dab5d7b33" exitCode=255 Mar 11 13:23:14 crc kubenswrapper[4923]: I0311 13:23:14.257299 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"3bf1a33b60de01496d3f642c6b988be5ce449a4c1a7b02919a4af52dab5d7b33"} Mar 11 13:23:14 crc kubenswrapper[4923]: I0311 13:23:14.257387 4923 scope.go:117] "RemoveContainer" containerID="0de3d701084f9234db9bb4b76d24ce0d3d40c90cd3422ed49b522207d498de90" Mar 11 13:23:14 crc kubenswrapper[4923]: I0311 13:23:14.257605 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:14 crc kubenswrapper[4923]: I0311 13:23:14.258846 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:14 crc kubenswrapper[4923]: I0311 13:23:14.258884 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:14 crc kubenswrapper[4923]: I0311 13:23:14.258895 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:14 crc kubenswrapper[4923]: I0311 13:23:14.259460 4923 scope.go:117] "RemoveContainer" containerID="3bf1a33b60de01496d3f642c6b988be5ce449a4c1a7b02919a4af52dab5d7b33" Mar 11 13:23:14 crc kubenswrapper[4923]: E0311 13:23:14.259620 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 11 13:23:14 crc kubenswrapper[4923]: I0311 13:23:14.670372 4923 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 11 13:23:14 crc kubenswrapper[4923]: I0311 13:23:14.670482 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 11 13:23:14 crc kubenswrapper[4923]: E0311 13:23:14.765761 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 11 13:23:14 crc kubenswrapper[4923]: I0311 13:23:14.767795 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:14 crc kubenswrapper[4923]: I0311 13:23:14.769727 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:14 crc kubenswrapper[4923]: I0311 13:23:14.769803 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:14 crc kubenswrapper[4923]: I0311 13:23:14.769827 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:14 crc kubenswrapper[4923]: I0311 13:23:14.769862 4923 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 11 13:23:14 crc kubenswrapper[4923]: E0311 13:23:14.775169 4923 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 11 13:23:14 crc kubenswrapper[4923]: I0311 13:23:14.949091 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:15 crc kubenswrapper[4923]: I0311 13:23:15.264732 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 11 13:23:15 crc kubenswrapper[4923]: I0311 13:23:15.269542 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:15 crc kubenswrapper[4923]: I0311 13:23:15.271271 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:15 crc kubenswrapper[4923]: I0311 13:23:15.271389 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:15 crc kubenswrapper[4923]: I0311 13:23:15.271413 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:15 crc kubenswrapper[4923]: I0311 13:23:15.272432 4923 scope.go:117] "RemoveContainer" containerID="3bf1a33b60de01496d3f642c6b988be5ce449a4c1a7b02919a4af52dab5d7b33" Mar 11 13:23:15 crc kubenswrapper[4923]: E0311 13:23:15.272776 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 11 13:23:15 crc kubenswrapper[4923]: I0311 13:23:15.947538 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:16 crc kubenswrapper[4923]: W0311 13:23:16.688540 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 11 13:23:16 crc kubenswrapper[4923]: E0311 13:23:16.688619 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 11 13:23:16 crc kubenswrapper[4923]: I0311 13:23:16.945476 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.365825 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2893102e53 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:32.938491475 +0000 UTC m=+0.620393189,LastTimestamp:2026-03-11 13:22:32.938491475 +0000 UTC m=+0.620393189,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.373865 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896be84c8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.00024852 +0000 UTC m=+0.682150304,LastTimestamp:2026-03-11 13:22:33.00024852 +0000 UTC m=+0.682150304,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.382118 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896c0cd84 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.000398212 +0000 UTC m=+0.682299966,LastTimestamp:2026-03-11 13:22:33.000398212 +0000 UTC m=+0.682299966,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.389887 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896c11a66 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.000417894 +0000 UTC m=+0.682319648,LastTimestamp:2026-03-11 13:22:33.000417894 +0000 UTC m=+0.682319648,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.395205 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc289c4d9382 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.093510018 +0000 UTC m=+0.775411752,LastTimestamp:2026-03-11 13:22:33.093510018 +0000 UTC m=+0.775411752,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.400719 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bcc2896be84c8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896be84c8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.00024852 +0000 UTC m=+0.682150304,LastTimestamp:2026-03-11 13:22:33.125026931 +0000 UTC m=+0.806928645,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.405578 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bcc2896c0cd84\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896c0cd84 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.000398212 +0000 UTC m=+0.682299966,LastTimestamp:2026-03-11 13:22:33.125055233 +0000 UTC m=+0.806956947,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.414450 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bcc2896c11a66\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896c11a66 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.000417894 +0000 UTC m=+0.682319648,LastTimestamp:2026-03-11 13:22:33.125067994 +0000 UTC m=+0.806969708,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.416892 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bcc2896be84c8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896be84c8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.00024852 +0000 UTC m=+0.682150304,LastTimestamp:2026-03-11 13:22:33.126425372 +0000 UTC m=+0.808327096,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.424434 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bcc2896c0cd84\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896c0cd84 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.000398212 +0000 UTC m=+0.682299966,LastTimestamp:2026-03-11 13:22:33.126439853 +0000 UTC m=+0.808341577,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.431836 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bcc2896c11a66\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896c11a66 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.000417894 +0000 UTC m=+0.682319648,LastTimestamp:2026-03-11 13:22:33.126451774 +0000 UTC m=+0.808353498,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.437582 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bcc2896be84c8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896be84c8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.00024852 +0000 UTC m=+0.682150304,LastTimestamp:2026-03-11 13:22:33.126522879 +0000 UTC m=+0.808424593,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.444536 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bcc2896c0cd84\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896c0cd84 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.000398212 +0000 UTC m=+0.682299966,LastTimestamp:2026-03-11 13:22:33.126548151 +0000 UTC m=+0.808449865,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.449204 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bcc2896c11a66\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896c11a66 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.000417894 +0000 UTC m=+0.682319648,LastTimestamp:2026-03-11 13:22:33.126557382 +0000 UTC m=+0.808459096,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.456521 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bcc2896be84c8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896be84c8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.00024852 +0000 UTC m=+0.682150304,LastTimestamp:2026-03-11 13:22:33.127334803 +0000 UTC m=+0.809236517,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.463808 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bcc2896c0cd84\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896c0cd84 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.000398212 +0000 UTC m=+0.682299966,LastTimestamp:2026-03-11 13:22:33.127375097 +0000 UTC m=+0.809276811,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.470596 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bcc2896c11a66\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896c11a66 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.000417894 +0000 UTC m=+0.682319648,LastTimestamp:2026-03-11 13:22:33.127387528 +0000 UTC m=+0.809289242,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.477603 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bcc2896be84c8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896be84c8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.00024852 +0000 UTC m=+0.682150304,LastTimestamp:2026-03-11 13:22:33.128363505 +0000 UTC m=+0.810265239,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.484389 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bcc2896c0cd84\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896c0cd84 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.000398212 +0000 UTC m=+0.682299966,LastTimestamp:2026-03-11 13:22:33.128387986 +0000 UTC m=+0.810289710,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.490531 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bcc2896c11a66\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896c11a66 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.000417894 +0000 UTC m=+0.682319648,LastTimestamp:2026-03-11 13:22:33.128401217 +0000 UTC m=+0.810302951,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.496564 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bcc2896be84c8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896be84c8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.00024852 +0000 UTC m=+0.682150304,LastTimestamp:2026-03-11 13:22:33.12867892 +0000 UTC m=+0.810580674,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.502540 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bcc2896c0cd84\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896c0cd84 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.000398212 +0000 UTC m=+0.682299966,LastTimestamp:2026-03-11 13:22:33.128710133 +0000 UTC m=+0.810611887,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.509120 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bcc2896c11a66\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896c11a66 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.000417894 +0000 UTC m=+0.682319648,LastTimestamp:2026-03-11 13:22:33.128728444 +0000 UTC m=+0.810630188,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.517782 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bcc2896be84c8\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896be84c8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.00024852 +0000 UTC m=+0.682150304,LastTimestamp:2026-03-11 13:22:33.129055 +0000 UTC m=+0.810956714,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.521742 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bcc2896c0cd84\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bcc2896c0cd84 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.000398212 +0000 UTC m=+0.682299966,LastTimestamp:2026-03-11 13:22:33.129072121 +0000 UTC m=+0.810973835,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.525607 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189bcc28b6ead133 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.540022579 +0000 UTC m=+1.221924303,LastTimestamp:2026-03-11 13:22:33.540022579 +0000 UTC m=+1.221924303,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.533135 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bcc28b7339c97 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.544793239 +0000 UTC m=+1.226694983,LastTimestamp:2026-03-11 13:22:33.544793239 +0000 UTC m=+1.226694983,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.540152 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bcc28b78e2774 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.550727028 +0000 UTC m=+1.232628782,LastTimestamp:2026-03-11 13:22:33.550727028 +0000 UTC m=+1.232628782,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.546946 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc28b898daaf openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.568205487 +0000 UTC m=+1.250107211,LastTimestamp:2026-03-11 13:22:33.568205487 +0000 UTC m=+1.250107211,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: I0311 13:23:17.554751 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:23:17 crc kubenswrapper[4923]: I0311 13:23:17.554951 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.554839 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bcc28b91da111 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:33.576907025 +0000 UTC m=+1.258808779,LastTimestamp:2026-03-11 13:22:33.576907025 +0000 UTC m=+1.258808779,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: I0311 13:23:17.556621 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:17 crc kubenswrapper[4923]: I0311 13:23:17.556683 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:17 crc kubenswrapper[4923]: I0311 13:23:17.556698 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:17 crc kubenswrapper[4923]: I0311 13:23:17.557543 4923 scope.go:117] "RemoveContainer" containerID="3bf1a33b60de01496d3f642c6b988be5ce449a4c1a7b02919a4af52dab5d7b33" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.557801 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.560895 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bcc28d99ef448 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:34.122253384 +0000 UTC m=+1.804155118,LastTimestamp:2026-03-11 13:22:34.122253384 +0000 UTC m=+1.804155118,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.568151 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bcc28d99fad19 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:34.122300697 +0000 UTC m=+1.804202411,LastTimestamp:2026-03-11 13:22:34.122300697 +0000 UTC m=+1.804202411,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.573384 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bcc28d9bade2f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:34.124082735 +0000 UTC m=+1.805984449,LastTimestamp:2026-03-11 13:22:34.124082735 +0000 UTC m=+1.805984449,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.580580 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc28da255fc7 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:34.131062727 +0000 UTC m=+1.812964441,LastTimestamp:2026-03-11 13:22:34.131062727 +0000 UTC m=+1.812964441,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.587636 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189bcc28da365a05 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:34.132175365 +0000 UTC m=+1.814077079,LastTimestamp:2026-03-11 13:22:34.132175365 +0000 UTC m=+1.814077079,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.594590 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bcc28da40ccad openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:34.132860077 +0000 UTC m=+1.814761791,LastTimestamp:2026-03-11 13:22:34.132860077 +0000 UTC m=+1.814761791,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.599865 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bcc28da48a348 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:34.133373768 +0000 UTC m=+1.815275482,LastTimestamp:2026-03-11 13:22:34.133373768 +0000 UTC m=+1.815275482,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.606803 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bcc28da5ccf58 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:34.134695768 +0000 UTC m=+1.816597482,LastTimestamp:2026-03-11 13:22:34.134695768 +0000 UTC m=+1.816597482,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.611989 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bcc28da61c116 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:34.135019798 +0000 UTC m=+1.816921512,LastTimestamp:2026-03-11 13:22:34.135019798 +0000 UTC m=+1.816921512,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.619050 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc28db103ffe openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:34.14645555 +0000 UTC m=+1.828357264,LastTimestamp:2026-03-11 13:22:34.14645555 +0000 UTC m=+1.828357264,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.625023 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189bcc28db5e160c openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:34.15155662 +0000 UTC m=+1.833458334,LastTimestamp:2026-03-11 13:22:34.15155662 +0000 UTC m=+1.833458334,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.631114 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bcc28eb744214 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:34.42144514 +0000 UTC m=+2.103346884,LastTimestamp:2026-03-11 13:22:34.42144514 +0000 UTC m=+2.103346884,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.635464 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bcc28ec05c098 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:34.430980248 +0000 UTC m=+2.112881962,LastTimestamp:2026-03-11 13:22:34.430980248 +0000 UTC m=+2.112881962,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.642059 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bcc28ec1b7592 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:34.432402834 +0000 UTC m=+2.114304548,LastTimestamp:2026-03-11 13:22:34.432402834 +0000 UTC m=+2.114304548,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.648386 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bcc28f8bff26f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:34.644509295 +0000 UTC m=+2.326411029,LastTimestamp:2026-03-11 13:22:34.644509295 +0000 UTC m=+2.326411029,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.655935 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bcc28f9b5d3ed openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:34.660623341 +0000 UTC m=+2.342525085,LastTimestamp:2026-03-11 13:22:34.660623341 +0000 UTC m=+2.342525085,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.662694 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bcc28f9d22d29 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:34.662481193 +0000 UTC m=+2.344382987,LastTimestamp:2026-03-11 13:22:34.662481193 +0000 UTC m=+2.344382987,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.667579 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bcc29058813fc openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:34.858951676 +0000 UTC m=+2.540853430,LastTimestamp:2026-03-11 13:22:34.858951676 +0000 UTC m=+2.540853430,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.671427 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bcc2906222638 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:34.869048888 +0000 UTC m=+2.550950602,LastTimestamp:2026-03-11 13:22:34.869048888 +0000 UTC m=+2.550950602,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.678976 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189bcc2910d8b04b openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.048783947 +0000 UTC m=+2.730685702,LastTimestamp:2026-03-11 13:22:35.048783947 +0000 UTC m=+2.730685702,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.684824 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bcc2910f16856 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.050403926 +0000 UTC m=+2.732305630,LastTimestamp:2026-03-11 13:22:35.050403926 +0000 UTC m=+2.732305630,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.691336 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc291142527a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.055706746 +0000 UTC m=+2.737608500,LastTimestamp:2026-03-11 13:22:35.055706746 +0000 UTC m=+2.737608500,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.699270 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bcc2911a6942d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.062277165 +0000 UTC m=+2.744178919,LastTimestamp:2026-03-11 13:22:35.062277165 +0000 UTC m=+2.744178919,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.705883 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bcc2922f6afcd openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.352739789 +0000 UTC m=+3.034641543,LastTimestamp:2026-03-11 13:22:35.352739789 +0000 UTC m=+3.034641543,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.713007 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189bcc292371d4cd openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.360810189 +0000 UTC m=+3.042711903,LastTimestamp:2026-03-11 13:22:35.360810189 +0000 UTC m=+3.042711903,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.719121 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bcc292374423c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.360969276 +0000 UTC m=+3.042870980,LastTimestamp:2026-03-11 13:22:35.360969276 +0000 UTC m=+3.042870980,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.729050 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc2923758e52 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.36105429 +0000 UTC m=+3.042956004,LastTimestamp:2026-03-11 13:22:35.36105429 +0000 UTC m=+3.042956004,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.735728 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bcc2924069a4e openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.370560078 +0000 UTC m=+3.052461802,LastTimestamp:2026-03-11 13:22:35.370560078 +0000 UTC m=+3.052461802,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.741974 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc29244f1a55 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.375311445 +0000 UTC m=+3.057213189,LastTimestamp:2026-03-11 13:22:35.375311445 +0000 UTC m=+3.057213189,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.752448 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bcc292455a95c openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.375741276 +0000 UTC m=+3.057642990,LastTimestamp:2026-03-11 13:22:35.375741276 +0000 UTC m=+3.057642990,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.759919 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bcc292477157e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.377931646 +0000 UTC m=+3.059833390,LastTimestamp:2026-03-11 13:22:35.377931646 +0000 UTC m=+3.059833390,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.767594 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bcc292492012d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.379695917 +0000 UTC m=+3.061597671,LastTimestamp:2026-03-11 13:22:35.379695917 +0000 UTC m=+3.061597671,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.772585 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189bcc292552dffc openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.392335868 +0000 UTC m=+3.074237622,LastTimestamp:2026-03-11 13:22:35.392335868 +0000 UTC m=+3.074237622,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.779302 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bcc293033ad9a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.57484073 +0000 UTC m=+3.256742454,LastTimestamp:2026-03-11 13:22:35.57484073 +0000 UTC m=+3.256742454,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.785782 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bcc29305ebb8e openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.57766235 +0000 UTC m=+3.259564074,LastTimestamp:2026-03-11 13:22:35.57766235 +0000 UTC m=+3.259564074,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.793464 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bcc293145c9cf openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.592804815 +0000 UTC m=+3.274706539,LastTimestamp:2026-03-11 13:22:35.592804815 +0000 UTC m=+3.274706539,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.801564 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bcc2931542982 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.593746818 +0000 UTC m=+3.275648542,LastTimestamp:2026-03-11 13:22:35.593746818 +0000 UTC m=+3.275648542,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.808315 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bcc29315cdf39 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.594317625 +0000 UTC m=+3.276219399,LastTimestamp:2026-03-11 13:22:35.594317625 +0000 UTC m=+3.276219399,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.815466 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bcc2931617bd2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.594619858 +0000 UTC m=+3.276521582,LastTimestamp:2026-03-11 13:22:35.594619858 +0000 UTC m=+3.276521582,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.822300 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bcc293f83c649 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.831748169 +0000 UTC m=+3.513649893,LastTimestamp:2026-03-11 13:22:35.831748169 +0000 UTC m=+3.513649893,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.827922 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bcc293fb69177 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.835076983 +0000 UTC m=+3.516978707,LastTimestamp:2026-03-11 13:22:35.835076983 +0000 UTC m=+3.516978707,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.833646 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bcc294056ad20 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.845569824 +0000 UTC m=+3.527471548,LastTimestamp:2026-03-11 13:22:35.845569824 +0000 UTC m=+3.527471548,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.836845 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bcc2940c0cbb0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.852524464 +0000 UTC m=+3.534426178,LastTimestamp:2026-03-11 13:22:35.852524464 +0000 UTC m=+3.534426178,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.842152 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bcc2940d2b6e3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:35.853698787 +0000 UTC m=+3.535600511,LastTimestamp:2026-03-11 13:22:35.853698787 +0000 UTC m=+3.535600511,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.844281 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bcc294c6c2484 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:36.048303236 +0000 UTC m=+3.730204960,LastTimestamp:2026-03-11 13:22:36.048303236 +0000 UTC m=+3.730204960,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.847803 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bcc294d3fc651 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:36.062172753 +0000 UTC m=+3.744074467,LastTimestamp:2026-03-11 13:22:36.062172753 +0000 UTC m=+3.744074467,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.852287 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bcc294d6400f6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:36.064547062 +0000 UTC m=+3.746448816,LastTimestamp:2026-03-11 13:22:36.064547062 +0000 UTC m=+3.746448816,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.855658 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc294e9d47a5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:36.085077925 +0000 UTC m=+3.766979649,LastTimestamp:2026-03-11 13:22:36.085077925 +0000 UTC m=+3.766979649,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.862683 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bcc295a2184a5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:36.278293669 +0000 UTC m=+3.960195383,LastTimestamp:2026-03-11 13:22:36.278293669 +0000 UTC m=+3.960195383,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.869005 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc295a345441 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:36.279526465 +0000 UTC m=+3.961428179,LastTimestamp:2026-03-11 13:22:36.279526465 +0000 UTC m=+3.961428179,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.874545 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bcc295ae2f26c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:36.29097022 +0000 UTC m=+3.972871934,LastTimestamp:2026-03-11 13:22:36.29097022 +0000 UTC m=+3.972871934,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.879692 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc295b2ee059 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:36.295946329 +0000 UTC m=+3.977848043,LastTimestamp:2026-03-11 13:22:36.295946329 +0000 UTC m=+3.977848043,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.886866 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc298a7f584b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:37.089749067 +0000 UTC m=+4.771650821,LastTimestamp:2026-03-11 13:22:37.089749067 +0000 UTC m=+4.771650821,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.893761 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc299b444860 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:37.37109104 +0000 UTC m=+5.052992794,LastTimestamp:2026-03-11 13:22:37.37109104 +0000 UTC m=+5.052992794,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.900157 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc299c1fec1d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:37.385485341 +0000 UTC m=+5.067387095,LastTimestamp:2026-03-11 13:22:37.385485341 +0000 UTC m=+5.067387095,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.907106 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc299c3e5b69 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:37.387479913 +0000 UTC m=+5.069381677,LastTimestamp:2026-03-11 13:22:37.387479913 +0000 UTC m=+5.069381677,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.913334 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc29abc2f524 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:37.64782826 +0000 UTC m=+5.329730014,LastTimestamp:2026-03-11 13:22:37.64782826 +0000 UTC m=+5.329730014,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.919764 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc29ac751ade openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:37.659503326 +0000 UTC m=+5.341405080,LastTimestamp:2026-03-11 13:22:37.659503326 +0000 UTC m=+5.341405080,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.926241 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc29ac8b9603 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:37.660976643 +0000 UTC m=+5.342878387,LastTimestamp:2026-03-11 13:22:37.660976643 +0000 UTC m=+5.342878387,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.932968 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc29bde1fe3d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:37.951852093 +0000 UTC m=+5.633753847,LastTimestamp:2026-03-11 13:22:37.951852093 +0000 UTC m=+5.633753847,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.940272 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc29beae08a2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:37.965224098 +0000 UTC m=+5.647125852,LastTimestamp:2026-03-11 13:22:37.965224098 +0000 UTC m=+5.647125852,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: I0311 13:23:17.949456 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.949418 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc29becef287 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:37.967381127 +0000 UTC m=+5.649282871,LastTimestamp:2026-03-11 13:22:37.967381127 +0000 UTC m=+5.649282871,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.957000 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc29ccf3c1f1 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:38.204674545 +0000 UTC m=+5.886576289,LastTimestamp:2026-03-11 13:22:38.204674545 +0000 UTC m=+5.886576289,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.966129 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc29ce029271 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:38.222422641 +0000 UTC m=+5.904324385,LastTimestamp:2026-03-11 13:22:38.222422641 +0000 UTC m=+5.904324385,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.973317 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc29ce18047f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:38.223828095 +0000 UTC m=+5.905729839,LastTimestamp:2026-03-11 13:22:38.223828095 +0000 UTC m=+5.905729839,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.980012 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc29deaccb1a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:38.502013722 +0000 UTC m=+6.183915466,LastTimestamp:2026-03-11 13:22:38.502013722 +0000 UTC m=+6.183915466,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.986170 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bcc29dfe47ab2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:38.52244037 +0000 UTC m=+6.204342114,LastTimestamp:2026-03-11 13:22:38.52244037 +0000 UTC m=+6.204342114,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:17 crc kubenswrapper[4923]: E0311 13:23:17.997286 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 11 13:23:17 crc kubenswrapper[4923]: &Event{ObjectMeta:{kube-controller-manager-crc.189bcc2b4e49f35a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Mar 11 13:23:17 crc kubenswrapper[4923]: body: Mar 11 13:23:17 crc kubenswrapper[4923]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:44.66955145 +0000 UTC m=+12.351453164,LastTimestamp:2026-03-11 13:22:44.66955145 +0000 UTC m=+12.351453164,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 11 13:23:17 crc kubenswrapper[4923]: > Mar 11 13:23:18 crc kubenswrapper[4923]: E0311 13:23:18.004437 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bcc2b4e4abb16 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:44.669602582 +0000 UTC m=+12.351504296,LastTimestamp:2026-03-11 13:22:44.669602582 +0000 UTC m=+12.351504296,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:18 crc kubenswrapper[4923]: E0311 13:23:18.012122 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 11 13:23:18 crc kubenswrapper[4923]: &Event{ObjectMeta:{kube-apiserver-crc.189bcc2bedddf5f7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 11 13:23:18 crc kubenswrapper[4923]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 11 13:23:18 crc kubenswrapper[4923]: Mar 11 13:23:18 crc kubenswrapper[4923]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:47.346828791 +0000 UTC m=+15.028730515,LastTimestamp:2026-03-11 13:22:47.346828791 +0000 UTC m=+15.028730515,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 11 13:23:18 crc kubenswrapper[4923]: > Mar 11 13:23:18 crc kubenswrapper[4923]: E0311 13:23:18.019176 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bcc2bedde906c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:47.346868332 +0000 UTC m=+15.028770056,LastTimestamp:2026-03-11 13:22:47.346868332 +0000 UTC m=+15.028770056,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:18 crc kubenswrapper[4923]: E0311 13:23:18.026807 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189bcc2bedddf5f7\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 11 13:23:18 crc kubenswrapper[4923]: &Event{ObjectMeta:{kube-apiserver-crc.189bcc2bedddf5f7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 11 13:23:18 crc kubenswrapper[4923]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 11 13:23:18 crc kubenswrapper[4923]: Mar 11 13:23:18 crc kubenswrapper[4923]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:47.346828791 +0000 UTC m=+15.028730515,LastTimestamp:2026-03-11 13:22:47.352207499 +0000 UTC m=+15.034109223,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 11 13:23:18 crc kubenswrapper[4923]: > Mar 11 13:23:18 crc kubenswrapper[4923]: E0311 13:23:18.035206 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189bcc2bedde906c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bcc2bedde906c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:47.346868332 +0000 UTC m=+15.028770056,LastTimestamp:2026-03-11 13:22:47.35223263 +0000 UTC m=+15.034134354,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:18 crc kubenswrapper[4923]: E0311 13:23:18.041202 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 11 13:23:18 crc kubenswrapper[4923]: &Event{ObjectMeta:{kube-apiserver-crc.189bcc2bfa599ac4 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:ProbeError,Message:Liveness probe error: Get "https://192.168.126.11:17697/healthz": dial tcp 192.168.126.11:17697: connect: connection refused Mar 11 13:23:18 crc kubenswrapper[4923]: body: Mar 11 13:23:18 crc kubenswrapper[4923]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:47.5562585 +0000 UTC m=+15.238160254,LastTimestamp:2026-03-11 13:22:47.5562585 +0000 UTC m=+15.238160254,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 11 13:23:18 crc kubenswrapper[4923]: > Mar 11 13:23:18 crc kubenswrapper[4923]: E0311 13:23:18.045764 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bcc2bfa5b5957 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Unhealthy,Message:Liveness probe failed: Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:47.556372823 +0000 UTC m=+15.238274577,LastTimestamp:2026-03-11 13:22:47.556372823 +0000 UTC m=+15.238274577,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:18 crc kubenswrapper[4923]: E0311 13:23:18.049415 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189bcc294d6400f6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bcc294d6400f6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:36.064547062 +0000 UTC m=+3.746448816,LastTimestamp:2026-03-11 13:22:48.144156043 +0000 UTC m=+15.826057807,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:18 crc kubenswrapper[4923]: E0311 13:23:18.054634 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189bcc2b4e49f35a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 11 13:23:18 crc kubenswrapper[4923]: &Event{ObjectMeta:{kube-controller-manager-crc.189bcc2b4e49f35a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Mar 11 13:23:18 crc kubenswrapper[4923]: body: Mar 11 13:23:18 crc kubenswrapper[4923]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:44.66955145 +0000 UTC m=+12.351453164,LastTimestamp:2026-03-11 13:22:54.670168123 +0000 UTC m=+22.352069877,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 11 13:23:18 crc kubenswrapper[4923]: > Mar 11 13:23:18 crc kubenswrapper[4923]: E0311 13:23:18.061678 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189bcc2b4e4abb16\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bcc2b4e4abb16 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:44.669602582 +0000 UTC m=+12.351504296,LastTimestamp:2026-03-11 13:22:54.670240155 +0000 UTC m=+22.352141899,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:18 crc kubenswrapper[4923]: E0311 13:23:18.070495 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 11 13:23:18 crc kubenswrapper[4923]: &Event{ObjectMeta:{kube-controller-manager-crc.189bcc2ff678271a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 11 13:23:18 crc kubenswrapper[4923]: body: Mar 11 13:23:18 crc kubenswrapper[4923]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:23:04.671020826 +0000 UTC m=+32.352922580,LastTimestamp:2026-03-11 13:23:04.671020826 +0000 UTC m=+32.352922580,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 11 13:23:18 crc kubenswrapper[4923]: > Mar 11 13:23:18 crc kubenswrapper[4923]: E0311 13:23:18.077458 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bcc2ff6798012 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:23:04.671109138 +0000 UTC m=+32.353010892,LastTimestamp:2026-03-11 13:23:04.671109138 +0000 UTC m=+32.353010892,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:18 crc kubenswrapper[4923]: E0311 13:23:18.084613 4923 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bcc2ff6d63ccb openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:23:04.677186763 +0000 UTC m=+32.359088517,LastTimestamp:2026-03-11 13:23:04.677186763 +0000 UTC m=+32.359088517,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:18 crc kubenswrapper[4923]: E0311 13:23:18.091475 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189bcc28da61c116\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bcc28da61c116 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:34.135019798 +0000 UTC m=+1.816921512,LastTimestamp:2026-03-11 13:23:04.799087523 +0000 UTC m=+32.480989237,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:18 crc kubenswrapper[4923]: E0311 13:23:18.099308 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189bcc28eb744214\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bcc28eb744214 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:34.42144514 +0000 UTC m=+2.103346884,LastTimestamp:2026-03-11 13:23:05.050028303 +0000 UTC m=+32.731930027,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:18 crc kubenswrapper[4923]: E0311 13:23:18.106258 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189bcc28ec05c098\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bcc28ec05c098 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:22:34.430980248 +0000 UTC m=+2.112881962,LastTimestamp:2026-03-11 13:23:05.062313268 +0000 UTC m=+32.744214992,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:18 crc kubenswrapper[4923]: E0311 13:23:18.115415 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189bcc2ff678271a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 11 13:23:18 crc kubenswrapper[4923]: &Event{ObjectMeta:{kube-controller-manager-crc.189bcc2ff678271a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 11 13:23:18 crc kubenswrapper[4923]: body: Mar 11 13:23:18 crc kubenswrapper[4923]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:23:04.671020826 +0000 UTC m=+32.352922580,LastTimestamp:2026-03-11 13:23:14.670444785 +0000 UTC m=+42.352346529,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 11 13:23:18 crc kubenswrapper[4923]: > Mar 11 13:23:18 crc kubenswrapper[4923]: E0311 13:23:18.122834 4923 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189bcc2ff6798012\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bcc2ff6798012 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:23:04.671109138 +0000 UTC m=+32.353010892,LastTimestamp:2026-03-11 13:23:14.670529618 +0000 UTC m=+42.352431372,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:23:18 crc kubenswrapper[4923]: I0311 13:23:18.947642 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:19 crc kubenswrapper[4923]: I0311 13:23:19.947650 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:20 crc kubenswrapper[4923]: I0311 13:23:20.949102 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:21 crc kubenswrapper[4923]: I0311 13:23:21.676524 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:23:21 crc kubenswrapper[4923]: I0311 13:23:21.676772 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:21 crc kubenswrapper[4923]: I0311 13:23:21.678394 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:21 crc kubenswrapper[4923]: I0311 13:23:21.678459 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:21 crc kubenswrapper[4923]: I0311 13:23:21.678487 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:21 crc kubenswrapper[4923]: I0311 13:23:21.683832 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:23:21 crc kubenswrapper[4923]: E0311 13:23:21.773611 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 11 13:23:21 crc kubenswrapper[4923]: I0311 13:23:21.775897 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:21 crc kubenswrapper[4923]: I0311 13:23:21.777584 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:21 crc kubenswrapper[4923]: I0311 13:23:21.777641 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:21 crc kubenswrapper[4923]: I0311 13:23:21.777696 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:21 crc kubenswrapper[4923]: I0311 13:23:21.777732 4923 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 11 13:23:21 crc kubenswrapper[4923]: E0311 13:23:21.784272 4923 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 11 13:23:21 crc kubenswrapper[4923]: I0311 13:23:21.948890 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:22 crc kubenswrapper[4923]: I0311 13:23:22.291876 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:22 crc kubenswrapper[4923]: I0311 13:23:22.293097 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:22 crc kubenswrapper[4923]: I0311 13:23:22.293245 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:22 crc kubenswrapper[4923]: I0311 13:23:22.293270 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:22 crc kubenswrapper[4923]: I0311 13:23:22.948237 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:23 crc kubenswrapper[4923]: E0311 13:23:23.099127 4923 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 11 13:23:23 crc kubenswrapper[4923]: W0311 13:23:23.177297 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 11 13:23:23 crc kubenswrapper[4923]: E0311 13:23:23.177415 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 11 13:23:23 crc kubenswrapper[4923]: I0311 13:23:23.950424 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:24 crc kubenswrapper[4923]: I0311 13:23:24.945265 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:25 crc kubenswrapper[4923]: I0311 13:23:25.948047 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:26 crc kubenswrapper[4923]: I0311 13:23:26.944615 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:27 crc kubenswrapper[4923]: I0311 13:23:27.043653 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 11 13:23:27 crc kubenswrapper[4923]: I0311 13:23:27.044247 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:27 crc kubenswrapper[4923]: I0311 13:23:27.045391 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:27 crc kubenswrapper[4923]: I0311 13:23:27.045514 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:27 crc kubenswrapper[4923]: I0311 13:23:27.045611 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:27 crc kubenswrapper[4923]: I0311 13:23:27.946436 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:28 crc kubenswrapper[4923]: E0311 13:23:28.780521 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 11 13:23:28 crc kubenswrapper[4923]: I0311 13:23:28.784421 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:28 crc kubenswrapper[4923]: I0311 13:23:28.786206 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:28 crc kubenswrapper[4923]: I0311 13:23:28.786272 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:28 crc kubenswrapper[4923]: I0311 13:23:28.786292 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:28 crc kubenswrapper[4923]: I0311 13:23:28.786373 4923 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 11 13:23:28 crc kubenswrapper[4923]: E0311 13:23:28.794853 4923 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 11 13:23:28 crc kubenswrapper[4923]: I0311 13:23:28.946014 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:29 crc kubenswrapper[4923]: I0311 13:23:29.021909 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:29 crc kubenswrapper[4923]: I0311 13:23:29.023621 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:29 crc kubenswrapper[4923]: I0311 13:23:29.023699 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:29 crc kubenswrapper[4923]: I0311 13:23:29.023728 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:29 crc kubenswrapper[4923]: I0311 13:23:29.024788 4923 scope.go:117] "RemoveContainer" containerID="3bf1a33b60de01496d3f642c6b988be5ce449a4c1a7b02919a4af52dab5d7b33" Mar 11 13:23:29 crc kubenswrapper[4923]: E0311 13:23:29.025153 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 11 13:23:29 crc kubenswrapper[4923]: I0311 13:23:29.941791 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:30 crc kubenswrapper[4923]: I0311 13:23:30.946950 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:31 crc kubenswrapper[4923]: I0311 13:23:31.947069 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:32 crc kubenswrapper[4923]: I0311 13:23:32.948595 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:33 crc kubenswrapper[4923]: E0311 13:23:33.099987 4923 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 11 13:23:33 crc kubenswrapper[4923]: I0311 13:23:33.946973 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:34 crc kubenswrapper[4923]: I0311 13:23:34.949721 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:35 crc kubenswrapper[4923]: E0311 13:23:35.789373 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 11 13:23:35 crc kubenswrapper[4923]: I0311 13:23:35.795869 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:35 crc kubenswrapper[4923]: I0311 13:23:35.798019 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:35 crc kubenswrapper[4923]: I0311 13:23:35.798376 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:35 crc kubenswrapper[4923]: I0311 13:23:35.798390 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:35 crc kubenswrapper[4923]: I0311 13:23:35.798418 4923 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 11 13:23:35 crc kubenswrapper[4923]: E0311 13:23:35.805584 4923 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 11 13:23:35 crc kubenswrapper[4923]: I0311 13:23:35.948670 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:36 crc kubenswrapper[4923]: I0311 13:23:36.946622 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 11 13:23:37 crc kubenswrapper[4923]: I0311 13:23:37.601579 4923 csr.go:261] certificate signing request csr-cwm9v is approved, waiting to be issued Mar 11 13:23:37 crc kubenswrapper[4923]: I0311 13:23:37.610751 4923 csr.go:257] certificate signing request csr-cwm9v is issued Mar 11 13:23:37 crc kubenswrapper[4923]: I0311 13:23:37.695097 4923 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 11 13:23:37 crc kubenswrapper[4923]: I0311 13:23:37.793325 4923 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 11 13:23:38 crc kubenswrapper[4923]: I0311 13:23:38.612678 4923 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2027-01-03 10:08:41.123096262 +0000 UTC Mar 11 13:23:38 crc kubenswrapper[4923]: I0311 13:23:38.612794 4923 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7148h45m2.510307698s for next certificate rotation Mar 11 13:23:40 crc kubenswrapper[4923]: I0311 13:23:40.021434 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:40 crc kubenswrapper[4923]: I0311 13:23:40.023036 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:40 crc kubenswrapper[4923]: I0311 13:23:40.023105 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:40 crc kubenswrapper[4923]: I0311 13:23:40.023129 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:40 crc kubenswrapper[4923]: I0311 13:23:40.024151 4923 scope.go:117] "RemoveContainer" containerID="3bf1a33b60de01496d3f642c6b988be5ce449a4c1a7b02919a4af52dab5d7b33" Mar 11 13:23:40 crc kubenswrapper[4923]: I0311 13:23:40.347515 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 11 13:23:40 crc kubenswrapper[4923]: I0311 13:23:40.349742 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9a9f78b1ff5dfa00fd3d80bafc5c131e5fd12de6d0556229087316aad81df459"} Mar 11 13:23:40 crc kubenswrapper[4923]: I0311 13:23:40.349951 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:40 crc kubenswrapper[4923]: I0311 13:23:40.351251 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:40 crc kubenswrapper[4923]: I0311 13:23:40.351327 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:40 crc kubenswrapper[4923]: I0311 13:23:40.351378 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:41 crc kubenswrapper[4923]: I0311 13:23:41.354821 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 11 13:23:41 crc kubenswrapper[4923]: I0311 13:23:41.355637 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 11 13:23:41 crc kubenswrapper[4923]: I0311 13:23:41.358216 4923 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9a9f78b1ff5dfa00fd3d80bafc5c131e5fd12de6d0556229087316aad81df459" exitCode=255 Mar 11 13:23:41 crc kubenswrapper[4923]: I0311 13:23:41.358259 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"9a9f78b1ff5dfa00fd3d80bafc5c131e5fd12de6d0556229087316aad81df459"} Mar 11 13:23:41 crc kubenswrapper[4923]: I0311 13:23:41.358338 4923 scope.go:117] "RemoveContainer" containerID="3bf1a33b60de01496d3f642c6b988be5ce449a4c1a7b02919a4af52dab5d7b33" Mar 11 13:23:41 crc kubenswrapper[4923]: I0311 13:23:41.358518 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:41 crc kubenswrapper[4923]: I0311 13:23:41.360272 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:41 crc kubenswrapper[4923]: I0311 13:23:41.360315 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:41 crc kubenswrapper[4923]: I0311 13:23:41.360339 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:41 crc kubenswrapper[4923]: I0311 13:23:41.361407 4923 scope.go:117] "RemoveContainer" containerID="9a9f78b1ff5dfa00fd3d80bafc5c131e5fd12de6d0556229087316aad81df459" Mar 11 13:23:41 crc kubenswrapper[4923]: E0311 13:23:41.361736 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.364116 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.806219 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.807857 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.807917 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.807942 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.808122 4923 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.820123 4923 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.820497 4923 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 11 13:23:42 crc kubenswrapper[4923]: E0311 13:23:42.820532 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.827864 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.827919 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.827939 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.827963 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.827982 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:23:42Z","lastTransitionTime":"2026-03-11T13:23:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:23:42 crc kubenswrapper[4923]: E0311 13:23:42.847282 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"77515c10-0352-4c6b-bba7-83e5133152ad\\\",\\\"systemUUID\\\":\\\"cf7f6afa-08d5-4a6d-91e6-356d0ec73635\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.856401 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.856453 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.856466 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.856486 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.856498 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:23:42Z","lastTransitionTime":"2026-03-11T13:23:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:23:42 crc kubenswrapper[4923]: E0311 13:23:42.869495 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"77515c10-0352-4c6b-bba7-83e5133152ad\\\",\\\"systemUUID\\\":\\\"cf7f6afa-08d5-4a6d-91e6-356d0ec73635\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.878895 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.878929 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.878938 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.878955 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.878966 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:23:42Z","lastTransitionTime":"2026-03-11T13:23:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:23:42 crc kubenswrapper[4923]: E0311 13:23:42.891215 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"77515c10-0352-4c6b-bba7-83e5133152ad\\\",\\\"systemUUID\\\":\\\"cf7f6afa-08d5-4a6d-91e6-356d0ec73635\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.900332 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.900413 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.900432 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.900454 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:23:42 crc kubenswrapper[4923]: I0311 13:23:42.900472 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:23:42Z","lastTransitionTime":"2026-03-11T13:23:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:23:42 crc kubenswrapper[4923]: E0311 13:23:42.915066 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"77515c10-0352-4c6b-bba7-83e5133152ad\\\",\\\"systemUUID\\\":\\\"cf7f6afa-08d5-4a6d-91e6-356d0ec73635\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 11 13:23:42 crc kubenswrapper[4923]: E0311 13:23:42.915262 4923 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 11 13:23:42 crc kubenswrapper[4923]: E0311 13:23:42.915288 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:43 crc kubenswrapper[4923]: E0311 13:23:43.015865 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:43 crc kubenswrapper[4923]: E0311 13:23:43.100643 4923 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 11 13:23:43 crc kubenswrapper[4923]: E0311 13:23:43.115979 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:43 crc kubenswrapper[4923]: E0311 13:23:43.216854 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:43 crc kubenswrapper[4923]: E0311 13:23:43.317549 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:43 crc kubenswrapper[4923]: E0311 13:23:43.418884 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:43 crc kubenswrapper[4923]: E0311 13:23:43.519062 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:43 crc kubenswrapper[4923]: E0311 13:23:43.619975 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:43 crc kubenswrapper[4923]: I0311 13:23:43.696569 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:23:43 crc kubenswrapper[4923]: I0311 13:23:43.696799 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:43 crc kubenswrapper[4923]: I0311 13:23:43.698461 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:43 crc kubenswrapper[4923]: I0311 13:23:43.698509 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:43 crc kubenswrapper[4923]: I0311 13:23:43.698525 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:43 crc kubenswrapper[4923]: I0311 13:23:43.700008 4923 scope.go:117] "RemoveContainer" containerID="9a9f78b1ff5dfa00fd3d80bafc5c131e5fd12de6d0556229087316aad81df459" Mar 11 13:23:43 crc kubenswrapper[4923]: E0311 13:23:43.700275 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 11 13:23:43 crc kubenswrapper[4923]: E0311 13:23:43.720966 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:43 crc kubenswrapper[4923]: E0311 13:23:43.822046 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:43 crc kubenswrapper[4923]: E0311 13:23:43.923222 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:44 crc kubenswrapper[4923]: E0311 13:23:44.023470 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:44 crc kubenswrapper[4923]: E0311 13:23:44.124371 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:44 crc kubenswrapper[4923]: E0311 13:23:44.224903 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:44 crc kubenswrapper[4923]: E0311 13:23:44.326073 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:44 crc kubenswrapper[4923]: E0311 13:23:44.426241 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:44 crc kubenswrapper[4923]: I0311 13:23:44.433563 4923 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 11 13:23:44 crc kubenswrapper[4923]: E0311 13:23:44.526673 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:44 crc kubenswrapper[4923]: E0311 13:23:44.627786 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:44 crc kubenswrapper[4923]: E0311 13:23:44.728171 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:44 crc kubenswrapper[4923]: E0311 13:23:44.829012 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:44 crc kubenswrapper[4923]: E0311 13:23:44.929764 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:45 crc kubenswrapper[4923]: E0311 13:23:45.030458 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:45 crc kubenswrapper[4923]: E0311 13:23:45.131446 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:45 crc kubenswrapper[4923]: E0311 13:23:45.232263 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:45 crc kubenswrapper[4923]: E0311 13:23:45.333395 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:45 crc kubenswrapper[4923]: E0311 13:23:45.433690 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:45 crc kubenswrapper[4923]: E0311 13:23:45.534804 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:45 crc kubenswrapper[4923]: E0311 13:23:45.636028 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:45 crc kubenswrapper[4923]: E0311 13:23:45.736206 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:45 crc kubenswrapper[4923]: E0311 13:23:45.836709 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:45 crc kubenswrapper[4923]: E0311 13:23:45.937513 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:46 crc kubenswrapper[4923]: E0311 13:23:46.038109 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:46 crc kubenswrapper[4923]: E0311 13:23:46.138505 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:46 crc kubenswrapper[4923]: E0311 13:23:46.238672 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:46 crc kubenswrapper[4923]: E0311 13:23:46.338942 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:46 crc kubenswrapper[4923]: E0311 13:23:46.439375 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:46 crc kubenswrapper[4923]: E0311 13:23:46.539827 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:46 crc kubenswrapper[4923]: E0311 13:23:46.640527 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:46 crc kubenswrapper[4923]: E0311 13:23:46.740704 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:46 crc kubenswrapper[4923]: E0311 13:23:46.841250 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:46 crc kubenswrapper[4923]: E0311 13:23:46.942209 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:47 crc kubenswrapper[4923]: E0311 13:23:47.043442 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:47 crc kubenswrapper[4923]: E0311 13:23:47.144437 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:47 crc kubenswrapper[4923]: E0311 13:23:47.244657 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:47 crc kubenswrapper[4923]: E0311 13:23:47.345247 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:47 crc kubenswrapper[4923]: E0311 13:23:47.445862 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:47 crc kubenswrapper[4923]: E0311 13:23:47.545993 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:47 crc kubenswrapper[4923]: I0311 13:23:47.554860 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:23:47 crc kubenswrapper[4923]: I0311 13:23:47.555048 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:23:47 crc kubenswrapper[4923]: I0311 13:23:47.556492 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:47 crc kubenswrapper[4923]: I0311 13:23:47.556552 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:47 crc kubenswrapper[4923]: I0311 13:23:47.556565 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:47 crc kubenswrapper[4923]: I0311 13:23:47.557277 4923 scope.go:117] "RemoveContainer" containerID="9a9f78b1ff5dfa00fd3d80bafc5c131e5fd12de6d0556229087316aad81df459" Mar 11 13:23:47 crc kubenswrapper[4923]: E0311 13:23:47.557545 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 11 13:23:47 crc kubenswrapper[4923]: E0311 13:23:47.646992 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:47 crc kubenswrapper[4923]: E0311 13:23:47.748142 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:47 crc kubenswrapper[4923]: E0311 13:23:47.849220 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:47 crc kubenswrapper[4923]: E0311 13:23:47.949870 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:48 crc kubenswrapper[4923]: E0311 13:23:48.050292 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:48 crc kubenswrapper[4923]: E0311 13:23:48.151259 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:48 crc kubenswrapper[4923]: E0311 13:23:48.252157 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:48 crc kubenswrapper[4923]: E0311 13:23:48.352670 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:48 crc kubenswrapper[4923]: E0311 13:23:48.453765 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:48 crc kubenswrapper[4923]: E0311 13:23:48.553984 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:48 crc kubenswrapper[4923]: E0311 13:23:48.654517 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:48 crc kubenswrapper[4923]: E0311 13:23:48.755586 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:48 crc kubenswrapper[4923]: E0311 13:23:48.855972 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:48 crc kubenswrapper[4923]: E0311 13:23:48.956716 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:49 crc kubenswrapper[4923]: E0311 13:23:49.057126 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:49 crc kubenswrapper[4923]: E0311 13:23:49.158529 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:49 crc kubenswrapper[4923]: E0311 13:23:49.259664 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:49 crc kubenswrapper[4923]: E0311 13:23:49.360030 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:49 crc kubenswrapper[4923]: E0311 13:23:49.460707 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:49 crc kubenswrapper[4923]: E0311 13:23:49.560841 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:49 crc kubenswrapper[4923]: E0311 13:23:49.661891 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:49 crc kubenswrapper[4923]: I0311 13:23:49.725893 4923 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 11 13:23:49 crc kubenswrapper[4923]: E0311 13:23:49.762656 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:49 crc kubenswrapper[4923]: E0311 13:23:49.863289 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:49 crc kubenswrapper[4923]: E0311 13:23:49.963813 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:50 crc kubenswrapper[4923]: E0311 13:23:50.064562 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:50 crc kubenswrapper[4923]: E0311 13:23:50.165003 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:50 crc kubenswrapper[4923]: E0311 13:23:50.265379 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:50 crc kubenswrapper[4923]: E0311 13:23:50.366178 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:50 crc kubenswrapper[4923]: E0311 13:23:50.467422 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:50 crc kubenswrapper[4923]: E0311 13:23:50.568265 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:50 crc kubenswrapper[4923]: E0311 13:23:50.668457 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:50 crc kubenswrapper[4923]: E0311 13:23:50.769418 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:50 crc kubenswrapper[4923]: E0311 13:23:50.869907 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:50 crc kubenswrapper[4923]: E0311 13:23:50.970859 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:51 crc kubenswrapper[4923]: E0311 13:23:51.072008 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:51 crc kubenswrapper[4923]: E0311 13:23:51.173095 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:51 crc kubenswrapper[4923]: E0311 13:23:51.273914 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:51 crc kubenswrapper[4923]: E0311 13:23:51.374485 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:51 crc kubenswrapper[4923]: E0311 13:23:51.474598 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:51 crc kubenswrapper[4923]: E0311 13:23:51.575759 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:51 crc kubenswrapper[4923]: E0311 13:23:51.676761 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:51 crc kubenswrapper[4923]: E0311 13:23:51.777617 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:51 crc kubenswrapper[4923]: E0311 13:23:51.878392 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:51 crc kubenswrapper[4923]: E0311 13:23:51.978570 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:52 crc kubenswrapper[4923]: E0311 13:23:52.079089 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:52 crc kubenswrapper[4923]: E0311 13:23:52.180240 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:52 crc kubenswrapper[4923]: E0311 13:23:52.280999 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:52 crc kubenswrapper[4923]: E0311 13:23:52.381585 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:52 crc kubenswrapper[4923]: E0311 13:23:52.482681 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:52 crc kubenswrapper[4923]: E0311 13:23:52.583116 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:52 crc kubenswrapper[4923]: E0311 13:23:52.684223 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:52 crc kubenswrapper[4923]: E0311 13:23:52.785151 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:52 crc kubenswrapper[4923]: E0311 13:23:52.885749 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:52 crc kubenswrapper[4923]: E0311 13:23:52.986540 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:53 crc kubenswrapper[4923]: E0311 13:23:53.087068 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:53 crc kubenswrapper[4923]: E0311 13:23:53.101418 4923 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 11 13:23:53 crc kubenswrapper[4923]: E0311 13:23:53.173168 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 11 13:23:53 crc kubenswrapper[4923]: I0311 13:23:53.178750 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:53 crc kubenswrapper[4923]: I0311 13:23:53.178814 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:53 crc kubenswrapper[4923]: I0311 13:23:53.178839 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:53 crc kubenswrapper[4923]: I0311 13:23:53.178870 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:23:53 crc kubenswrapper[4923]: I0311 13:23:53.178892 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:23:53Z","lastTransitionTime":"2026-03-11T13:23:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:23:53 crc kubenswrapper[4923]: E0311 13:23:53.195672 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"77515c10-0352-4c6b-bba7-83e5133152ad\\\",\\\"systemUUID\\\":\\\"cf7f6afa-08d5-4a6d-91e6-356d0ec73635\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 11 13:23:53 crc kubenswrapper[4923]: I0311 13:23:53.201459 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:53 crc kubenswrapper[4923]: I0311 13:23:53.201518 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:53 crc kubenswrapper[4923]: I0311 13:23:53.201533 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:53 crc kubenswrapper[4923]: I0311 13:23:53.201555 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:23:53 crc kubenswrapper[4923]: I0311 13:23:53.201572 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:23:53Z","lastTransitionTime":"2026-03-11T13:23:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:23:53 crc kubenswrapper[4923]: E0311 13:23:53.216766 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"77515c10-0352-4c6b-bba7-83e5133152ad\\\",\\\"systemUUID\\\":\\\"cf7f6afa-08d5-4a6d-91e6-356d0ec73635\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 11 13:23:53 crc kubenswrapper[4923]: I0311 13:23:53.221859 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:53 crc kubenswrapper[4923]: I0311 13:23:53.221927 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:53 crc kubenswrapper[4923]: I0311 13:23:53.221951 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:53 crc kubenswrapper[4923]: I0311 13:23:53.221982 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:23:53 crc kubenswrapper[4923]: I0311 13:23:53.222007 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:23:53Z","lastTransitionTime":"2026-03-11T13:23:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:23:53 crc kubenswrapper[4923]: E0311 13:23:53.238594 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"77515c10-0352-4c6b-bba7-83e5133152ad\\\",\\\"systemUUID\\\":\\\"cf7f6afa-08d5-4a6d-91e6-356d0ec73635\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 11 13:23:53 crc kubenswrapper[4923]: I0311 13:23:53.243220 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:23:53 crc kubenswrapper[4923]: I0311 13:23:53.243278 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:23:53 crc kubenswrapper[4923]: I0311 13:23:53.243297 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:23:53 crc kubenswrapper[4923]: I0311 13:23:53.243323 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:23:53 crc kubenswrapper[4923]: I0311 13:23:53.243369 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:23:53Z","lastTransitionTime":"2026-03-11T13:23:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:23:53 crc kubenswrapper[4923]: E0311 13:23:53.256825 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:23:53Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"77515c10-0352-4c6b-bba7-83e5133152ad\\\",\\\"systemUUID\\\":\\\"cf7f6afa-08d5-4a6d-91e6-356d0ec73635\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 11 13:23:53 crc kubenswrapper[4923]: E0311 13:23:53.257102 4923 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 11 13:23:53 crc kubenswrapper[4923]: E0311 13:23:53.257156 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:53 crc kubenswrapper[4923]: E0311 13:23:53.358142 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:53 crc kubenswrapper[4923]: E0311 13:23:53.458402 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:53 crc kubenswrapper[4923]: E0311 13:23:53.559555 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:53 crc kubenswrapper[4923]: E0311 13:23:53.660002 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:53 crc kubenswrapper[4923]: E0311 13:23:53.760654 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:53 crc kubenswrapper[4923]: E0311 13:23:53.861119 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:53 crc kubenswrapper[4923]: E0311 13:23:53.962045 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:54 crc kubenswrapper[4923]: E0311 13:23:54.063338 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:54 crc kubenswrapper[4923]: E0311 13:23:54.163644 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:54 crc kubenswrapper[4923]: E0311 13:23:54.264547 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:54 crc kubenswrapper[4923]: E0311 13:23:54.364905 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:54 crc kubenswrapper[4923]: E0311 13:23:54.465748 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:54 crc kubenswrapper[4923]: E0311 13:23:54.565984 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:54 crc kubenswrapper[4923]: E0311 13:23:54.666205 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:54 crc kubenswrapper[4923]: E0311 13:23:54.766444 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:54 crc kubenswrapper[4923]: E0311 13:23:54.867099 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:54 crc kubenswrapper[4923]: E0311 13:23:54.967912 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:55 crc kubenswrapper[4923]: E0311 13:23:55.069021 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:55 crc kubenswrapper[4923]: E0311 13:23:55.169935 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:55 crc kubenswrapper[4923]: I0311 13:23:55.190850 4923 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 11 13:23:55 crc kubenswrapper[4923]: E0311 13:23:55.270438 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:55 crc kubenswrapper[4923]: E0311 13:23:55.370917 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:55 crc kubenswrapper[4923]: E0311 13:23:55.471931 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:55 crc kubenswrapper[4923]: E0311 13:23:55.572715 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:55 crc kubenswrapper[4923]: E0311 13:23:55.673675 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:55 crc kubenswrapper[4923]: E0311 13:23:55.774075 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:55 crc kubenswrapper[4923]: E0311 13:23:55.874449 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:55 crc kubenswrapper[4923]: E0311 13:23:55.975126 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:56 crc kubenswrapper[4923]: E0311 13:23:56.075938 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:56 crc kubenswrapper[4923]: E0311 13:23:56.176442 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:56 crc kubenswrapper[4923]: E0311 13:23:56.277550 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:56 crc kubenswrapper[4923]: E0311 13:23:56.378455 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:56 crc kubenswrapper[4923]: E0311 13:23:56.478736 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:56 crc kubenswrapper[4923]: E0311 13:23:56.579125 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:56 crc kubenswrapper[4923]: E0311 13:23:56.679620 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:56 crc kubenswrapper[4923]: E0311 13:23:56.780288 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:56 crc kubenswrapper[4923]: E0311 13:23:56.880716 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:56 crc kubenswrapper[4923]: E0311 13:23:56.981684 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:57 crc kubenswrapper[4923]: E0311 13:23:57.082438 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:57 crc kubenswrapper[4923]: E0311 13:23:57.183223 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:57 crc kubenswrapper[4923]: E0311 13:23:57.283608 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:57 crc kubenswrapper[4923]: E0311 13:23:57.384757 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:57 crc kubenswrapper[4923]: E0311 13:23:57.485693 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:57 crc kubenswrapper[4923]: E0311 13:23:57.586289 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:57 crc kubenswrapper[4923]: E0311 13:23:57.687122 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:57 crc kubenswrapper[4923]: E0311 13:23:57.787783 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:57 crc kubenswrapper[4923]: E0311 13:23:57.888268 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:57 crc kubenswrapper[4923]: E0311 13:23:57.989568 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:58 crc kubenswrapper[4923]: E0311 13:23:58.090457 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:58 crc kubenswrapper[4923]: E0311 13:23:58.190952 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:58 crc kubenswrapper[4923]: E0311 13:23:58.292119 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:58 crc kubenswrapper[4923]: E0311 13:23:58.392826 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:58 crc kubenswrapper[4923]: E0311 13:23:58.493225 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:58 crc kubenswrapper[4923]: E0311 13:23:58.594282 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:58 crc kubenswrapper[4923]: E0311 13:23:58.695931 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:58 crc kubenswrapper[4923]: E0311 13:23:58.796515 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:58 crc kubenswrapper[4923]: E0311 13:23:58.897175 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:58 crc kubenswrapper[4923]: E0311 13:23:58.998274 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:59 crc kubenswrapper[4923]: E0311 13:23:59.098733 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:59 crc kubenswrapper[4923]: E0311 13:23:59.199887 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:59 crc kubenswrapper[4923]: E0311 13:23:59.300534 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:59 crc kubenswrapper[4923]: E0311 13:23:59.401626 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:59 crc kubenswrapper[4923]: E0311 13:23:59.502820 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:59 crc kubenswrapper[4923]: E0311 13:23:59.603766 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:59 crc kubenswrapper[4923]: E0311 13:23:59.704768 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:59 crc kubenswrapper[4923]: E0311 13:23:59.805538 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:23:59 crc kubenswrapper[4923]: E0311 13:23:59.906302 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:00 crc kubenswrapper[4923]: E0311 13:24:00.007223 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:00 crc kubenswrapper[4923]: E0311 13:24:00.108243 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:00 crc kubenswrapper[4923]: E0311 13:24:00.208557 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:00 crc kubenswrapper[4923]: E0311 13:24:00.308722 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:00 crc kubenswrapper[4923]: E0311 13:24:00.408950 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:00 crc kubenswrapper[4923]: E0311 13:24:00.509164 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:00 crc kubenswrapper[4923]: E0311 13:24:00.610309 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:00 crc kubenswrapper[4923]: E0311 13:24:00.711462 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:00 crc kubenswrapper[4923]: E0311 13:24:00.811898 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:00 crc kubenswrapper[4923]: E0311 13:24:00.912411 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:01 crc kubenswrapper[4923]: E0311 13:24:01.013156 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:01 crc kubenswrapper[4923]: I0311 13:24:01.021944 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:24:01 crc kubenswrapper[4923]: I0311 13:24:01.028533 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:01 crc kubenswrapper[4923]: I0311 13:24:01.028585 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:01 crc kubenswrapper[4923]: I0311 13:24:01.028600 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:01 crc kubenswrapper[4923]: I0311 13:24:01.030014 4923 scope.go:117] "RemoveContainer" containerID="9a9f78b1ff5dfa00fd3d80bafc5c131e5fd12de6d0556229087316aad81df459" Mar 11 13:24:01 crc kubenswrapper[4923]: E0311 13:24:01.031433 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 11 13:24:01 crc kubenswrapper[4923]: E0311 13:24:01.113544 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:01 crc kubenswrapper[4923]: E0311 13:24:01.213690 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:01 crc kubenswrapper[4923]: E0311 13:24:01.313885 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:01 crc kubenswrapper[4923]: E0311 13:24:01.414970 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:01 crc kubenswrapper[4923]: E0311 13:24:01.516234 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:01 crc kubenswrapper[4923]: E0311 13:24:01.617219 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:01 crc kubenswrapper[4923]: E0311 13:24:01.717686 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:01 crc kubenswrapper[4923]: E0311 13:24:01.818098 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:01 crc kubenswrapper[4923]: E0311 13:24:01.918470 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:02 crc kubenswrapper[4923]: E0311 13:24:02.018633 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:02 crc kubenswrapper[4923]: I0311 13:24:02.020910 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 11 13:24:02 crc kubenswrapper[4923]: I0311 13:24:02.022138 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:02 crc kubenswrapper[4923]: I0311 13:24:02.022180 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:02 crc kubenswrapper[4923]: I0311 13:24:02.022193 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:02 crc kubenswrapper[4923]: E0311 13:24:02.119277 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:02 crc kubenswrapper[4923]: E0311 13:24:02.220538 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:02 crc kubenswrapper[4923]: E0311 13:24:02.320664 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:02 crc kubenswrapper[4923]: E0311 13:24:02.421673 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:02 crc kubenswrapper[4923]: E0311 13:24:02.522482 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:02 crc kubenswrapper[4923]: E0311 13:24:02.623195 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:02 crc kubenswrapper[4923]: E0311 13:24:02.723419 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:02 crc kubenswrapper[4923]: E0311 13:24:02.823825 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:02 crc kubenswrapper[4923]: E0311 13:24:02.924866 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:03 crc kubenswrapper[4923]: E0311 13:24:03.024964 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:03 crc kubenswrapper[4923]: E0311 13:24:03.102040 4923 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 11 13:24:03 crc kubenswrapper[4923]: E0311 13:24:03.126060 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:03 crc kubenswrapper[4923]: E0311 13:24:03.226401 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:03 crc kubenswrapper[4923]: E0311 13:24:03.326564 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:03 crc kubenswrapper[4923]: E0311 13:24:03.427524 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:03 crc kubenswrapper[4923]: E0311 13:24:03.528034 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:03 crc kubenswrapper[4923]: E0311 13:24:03.546502 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 11 13:24:03 crc kubenswrapper[4923]: I0311 13:24:03.551788 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:03 crc kubenswrapper[4923]: I0311 13:24:03.551880 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:03 crc kubenswrapper[4923]: I0311 13:24:03.551904 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:03 crc kubenswrapper[4923]: I0311 13:24:03.551979 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:03 crc kubenswrapper[4923]: I0311 13:24:03.552003 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:03Z","lastTransitionTime":"2026-03-11T13:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:03 crc kubenswrapper[4923]: E0311 13:24:03.567936 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"77515c10-0352-4c6b-bba7-83e5133152ad\\\",\\\"systemUUID\\\":\\\"cf7f6afa-08d5-4a6d-91e6-356d0ec73635\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 11 13:24:03 crc kubenswrapper[4923]: I0311 13:24:03.572653 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:03 crc kubenswrapper[4923]: I0311 13:24:03.572694 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:03 crc kubenswrapper[4923]: I0311 13:24:03.572706 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:03 crc kubenswrapper[4923]: I0311 13:24:03.572723 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:03 crc kubenswrapper[4923]: I0311 13:24:03.572735 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:03Z","lastTransitionTime":"2026-03-11T13:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:03 crc kubenswrapper[4923]: E0311 13:24:03.583560 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"77515c10-0352-4c6b-bba7-83e5133152ad\\\",\\\"systemUUID\\\":\\\"cf7f6afa-08d5-4a6d-91e6-356d0ec73635\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 11 13:24:03 crc kubenswrapper[4923]: I0311 13:24:03.587636 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:03 crc kubenswrapper[4923]: I0311 13:24:03.587687 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:03 crc kubenswrapper[4923]: I0311 13:24:03.587698 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:03 crc kubenswrapper[4923]: I0311 13:24:03.587718 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:03 crc kubenswrapper[4923]: I0311 13:24:03.587736 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:03Z","lastTransitionTime":"2026-03-11T13:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:03 crc kubenswrapper[4923]: E0311 13:24:03.599543 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"77515c10-0352-4c6b-bba7-83e5133152ad\\\",\\\"systemUUID\\\":\\\"cf7f6afa-08d5-4a6d-91e6-356d0ec73635\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 11 13:24:03 crc kubenswrapper[4923]: I0311 13:24:03.604577 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:03 crc kubenswrapper[4923]: I0311 13:24:03.604642 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:03 crc kubenswrapper[4923]: I0311 13:24:03.604659 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:03 crc kubenswrapper[4923]: I0311 13:24:03.604681 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:03 crc kubenswrapper[4923]: I0311 13:24:03.604697 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:03Z","lastTransitionTime":"2026-03-11T13:24:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:03 crc kubenswrapper[4923]: E0311 13:24:03.620789 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"77515c10-0352-4c6b-bba7-83e5133152ad\\\",\\\"systemUUID\\\":\\\"cf7f6afa-08d5-4a6d-91e6-356d0ec73635\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 11 13:24:03 crc kubenswrapper[4923]: E0311 13:24:03.621024 4923 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 11 13:24:03 crc kubenswrapper[4923]: E0311 13:24:03.628687 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:03 crc kubenswrapper[4923]: E0311 13:24:03.729067 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:03 crc kubenswrapper[4923]: E0311 13:24:03.830123 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:03 crc kubenswrapper[4923]: E0311 13:24:03.930930 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:04 crc kubenswrapper[4923]: E0311 13:24:04.031311 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:04 crc kubenswrapper[4923]: E0311 13:24:04.131650 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:04 crc kubenswrapper[4923]: E0311 13:24:04.232737 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:04 crc kubenswrapper[4923]: E0311 13:24:04.333901 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:04 crc kubenswrapper[4923]: E0311 13:24:04.434555 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:04 crc kubenswrapper[4923]: E0311 13:24:04.535392 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:04 crc kubenswrapper[4923]: E0311 13:24:04.636112 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:04 crc kubenswrapper[4923]: E0311 13:24:04.737186 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:04 crc kubenswrapper[4923]: E0311 13:24:04.838267 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:04 crc kubenswrapper[4923]: E0311 13:24:04.938454 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:05 crc kubenswrapper[4923]: E0311 13:24:05.039088 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:05 crc kubenswrapper[4923]: E0311 13:24:05.140248 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:05 crc kubenswrapper[4923]: E0311 13:24:05.240576 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:05 crc kubenswrapper[4923]: E0311 13:24:05.340876 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:05 crc kubenswrapper[4923]: E0311 13:24:05.440987 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:05 crc kubenswrapper[4923]: E0311 13:24:05.542073 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:05 crc kubenswrapper[4923]: E0311 13:24:05.643232 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:05 crc kubenswrapper[4923]: E0311 13:24:05.744373 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:05 crc kubenswrapper[4923]: E0311 13:24:05.844987 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:05 crc kubenswrapper[4923]: E0311 13:24:05.945450 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:06 crc kubenswrapper[4923]: E0311 13:24:06.046598 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:06 crc kubenswrapper[4923]: E0311 13:24:06.146778 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:06 crc kubenswrapper[4923]: E0311 13:24:06.247293 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:06 crc kubenswrapper[4923]: E0311 13:24:06.347419 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:06 crc kubenswrapper[4923]: E0311 13:24:06.447967 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:06 crc kubenswrapper[4923]: E0311 13:24:06.548705 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:06 crc kubenswrapper[4923]: E0311 13:24:06.648878 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:06 crc kubenswrapper[4923]: E0311 13:24:06.749403 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:06 crc kubenswrapper[4923]: E0311 13:24:06.849934 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:06 crc kubenswrapper[4923]: E0311 13:24:06.950322 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:07 crc kubenswrapper[4923]: E0311 13:24:07.051050 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:07 crc kubenswrapper[4923]: E0311 13:24:07.151442 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:07 crc kubenswrapper[4923]: E0311 13:24:07.252251 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:07 crc kubenswrapper[4923]: E0311 13:24:07.352623 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:07 crc kubenswrapper[4923]: E0311 13:24:07.452791 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:07 crc kubenswrapper[4923]: E0311 13:24:07.553179 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:07 crc kubenswrapper[4923]: E0311 13:24:07.654280 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:07 crc kubenswrapper[4923]: E0311 13:24:07.755333 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:07 crc kubenswrapper[4923]: E0311 13:24:07.856444 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:07 crc kubenswrapper[4923]: E0311 13:24:07.956587 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:08 crc kubenswrapper[4923]: E0311 13:24:08.057730 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:08 crc kubenswrapper[4923]: E0311 13:24:08.157876 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:08 crc kubenswrapper[4923]: E0311 13:24:08.258653 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:08 crc kubenswrapper[4923]: E0311 13:24:08.359583 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:08 crc kubenswrapper[4923]: E0311 13:24:08.460439 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:08 crc kubenswrapper[4923]: E0311 13:24:08.561122 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:08 crc kubenswrapper[4923]: E0311 13:24:08.661370 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:08 crc kubenswrapper[4923]: E0311 13:24:08.762420 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:08 crc kubenswrapper[4923]: E0311 13:24:08.863502 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:08 crc kubenswrapper[4923]: E0311 13:24:08.964373 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:09 crc kubenswrapper[4923]: E0311 13:24:09.065178 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:09 crc kubenswrapper[4923]: E0311 13:24:09.165951 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:09 crc kubenswrapper[4923]: E0311 13:24:09.266584 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:09 crc kubenswrapper[4923]: E0311 13:24:09.366828 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.460872 4923 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.469565 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.469632 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.469654 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.469679 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.469696 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:09Z","lastTransitionTime":"2026-03-11T13:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.572974 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.573040 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.573062 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.573095 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.573118 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:09Z","lastTransitionTime":"2026-03-11T13:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.676524 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.676579 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.676601 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.676631 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.676652 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:09Z","lastTransitionTime":"2026-03-11T13:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.779960 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.780079 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.780107 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.780135 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.780154 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:09Z","lastTransitionTime":"2026-03-11T13:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.884297 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.884406 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.884426 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.884460 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.884480 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:09Z","lastTransitionTime":"2026-03-11T13:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.988285 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.988427 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.988453 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.988481 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.988499 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:09Z","lastTransitionTime":"2026-03-11T13:24:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.989152 4923 apiserver.go:52] "Watching apiserver" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.998152 4923 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.998833 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.999391 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.999553 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 11 13:24:09 crc kubenswrapper[4923]: I0311 13:24:09.999718 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.000262 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.000756 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.000998 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.001086 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.001192 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.000220 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.003470 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.003471 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.003905 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.004256 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.004398 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.004400 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.006663 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.006754 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.006940 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.048594 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.048960 4923 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.070392 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.091738 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.092485 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.092620 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.092703 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.092900 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.093034 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.092776 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.093208 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.093231 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.093261 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.093280 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:10Z","lastTransitionTime":"2026-03-11T13:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.093156 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.093596 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.093647 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.093716 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.093744 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.093771 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.093801 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.093827 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.093852 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.093880 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.093906 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.093930 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.093954 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.093978 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094000 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094022 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094044 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094073 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094104 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094135 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094140 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094138 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094148 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094164 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094297 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094335 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094442 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094464 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094502 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094504 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094533 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094540 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094617 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094664 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094702 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094740 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094774 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094801 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.093568 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094624 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094850 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.096141 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.096170 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.094979 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.095270 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.095297 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.095329 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.095463 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.095565 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.096522 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.095595 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.095898 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.095976 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.096653 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.096014 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.096031 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.096774 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.096820 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.097042 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.097089 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.097102 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.097126 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.097285 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.097326 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.097324 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.097520 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:10.59748724 +0000 UTC m=+98.279388984 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.097382 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.097602 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.097648 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.097686 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.097739 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.097775 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.097808 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.097841 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.097873 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.097906 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.097938 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.097952 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.097971 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098006 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098037 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098082 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098136 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098182 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098286 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098382 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098432 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098481 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098519 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098574 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098641 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098692 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098740 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098775 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098833 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098902 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098958 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099010 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099060 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099112 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099152 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099200 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099384 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099454 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099493 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099534 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099569 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098088 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099671 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099712 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099750 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099786 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099607 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098441 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098443 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099833 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098664 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099879 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.098935 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099484 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099820 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099622 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099590 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099954 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099984 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100010 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100137 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.099958 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100168 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100409 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100417 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100463 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100504 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100506 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100541 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100580 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100617 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100653 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100653 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100657 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100736 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100762 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100783 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100803 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100822 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100838 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100855 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100874 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100891 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100911 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100931 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100947 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100946 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100964 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.100989 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.101041 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.101089 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.101127 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.101445 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.101488 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.101542 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.101589 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.101627 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.101634 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.101685 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.101722 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.101674 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.102968 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.103439 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.103422 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.103847 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.104504 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.104555 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.104702 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.106283 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.104808 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.104827 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.104845 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.104913 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.105585 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.105570 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.106020 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.106116 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.106192 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.106423 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.106835 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.107170 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.107332 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.108083 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.108298 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.108573 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.108596 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.108428 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.108867 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.108883 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.109442 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.109397 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.109570 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.110593 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.112099 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.112284 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.112915 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.112998 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.113146 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.113527 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.114158 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.114535 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.115746 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.115803 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.115840 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.101755 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.116133 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.116235 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.116241 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.116321 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.116482 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.116651 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.116739 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.116758 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.116772 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.116805 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.116841 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.116841 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.116870 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.116902 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.116885 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.116935 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.117331 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.117447 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.117458 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.117474 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.117544 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.117549 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.101759 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.117633 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.117708 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.117780 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.117883 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.117944 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.117983 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118033 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118069 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118102 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118125 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118152 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118168 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118180 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118290 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118327 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118410 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118445 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118484 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118512 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118541 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118570 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118595 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118699 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118734 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118771 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118809 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118837 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118872 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118901 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118931 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118939 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118958 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.118989 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119019 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119044 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119074 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119102 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119129 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119158 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119167 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119186 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119217 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119240 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119268 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119296 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119327 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119382 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119415 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119538 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119571 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119580 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119597 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119605 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119801 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119843 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.119876 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.120145 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.120100 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.120227 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.120167 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.120635 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.120665 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.120691 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.120775 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.120852 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.120963 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.121069 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.121189 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.121839 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.125914 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.122255 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.122581 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.122610 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.122721 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.123096 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.124652 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.125133 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.125762 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.126251 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.125894 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.126424 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.120939 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.125104 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.127359 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.126291 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.127212 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.127781 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.128019 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.128504 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.128621 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.128662 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.128870 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.128937 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.128880 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.129637 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.129721 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.130390 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.130512 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.130650 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.130662 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.132203 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.132842 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.132897 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.134439 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.134179 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.134495 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.134192 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.134363 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.134334 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.134603 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.130200 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.134943 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.134982 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135015 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135048 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135074 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135104 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135185 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135221 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135261 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135299 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135333 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135384 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135414 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135444 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135476 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135508 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135535 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135566 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135597 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135623 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.136118 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.134672 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.136270 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.134756 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.134878 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135224 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135474 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135584 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135564 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.135864 4923 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.133178 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.135987 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.136457 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.136542 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.137380 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.137503 4923 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.138039 4923 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.138314 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-11 13:24:10.638272613 +0000 UTC m=+98.320174387 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.138511 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-11 13:24:10.638488648 +0000 UTC m=+98.320390402 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.138310 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.138730 4923 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.139145 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.139435 4923 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.139554 4923 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.139646 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.139730 4923 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.139827 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.139919 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.140296 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.140422 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.140513 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.140597 4923 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.140686 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.140771 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.140856 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.140933 4923 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.141014 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.141096 4923 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.141174 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.141251 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.141423 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.141522 4923 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.141610 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.141696 4923 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.141778 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.155916 4923 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.155948 4923 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.139968 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.155742 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.140599 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.147607 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.151117 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.154144 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.155008 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157411 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157488 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157507 4923 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157521 4923 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157534 4923 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157546 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157557 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157568 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157577 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157589 4923 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157599 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157613 4923 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157623 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157635 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157646 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157657 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157669 4923 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157682 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157695 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157704 4923 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157715 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157727 4923 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157737 4923 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157748 4923 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157759 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157768 4923 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157779 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157789 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157801 4923 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157815 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157830 4923 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157841 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157851 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157861 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157871 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157881 4923 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157891 4923 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157902 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157912 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157923 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157933 4923 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157945 4923 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157955 4923 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157965 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157974 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157983 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.157992 4923 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158002 4923 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158012 4923 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158021 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158030 4923 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158040 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158050 4923 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158061 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158071 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158080 4923 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158089 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158099 4923 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158109 4923 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.158104 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.158146 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.158246 4923 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.158361 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-11 13:24:10.65831476 +0000 UTC m=+98.340216474 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158118 4923 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158431 4923 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158452 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158467 4923 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158483 4923 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158499 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158515 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158532 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158549 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158566 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158580 4923 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158595 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158608 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158622 4923 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158636 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158653 4923 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158670 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158711 4923 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158726 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158742 4923 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158756 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158771 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158788 4923 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158802 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158818 4923 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158845 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158863 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158879 4923 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158894 4923 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158914 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158928 4923 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158942 4923 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158958 4923 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158974 4923 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.158989 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159005 4923 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159035 4923 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159052 4923 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159067 4923 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159057 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159084 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159082 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159185 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159206 4923 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159224 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159241 4923 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159256 4923 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159269 4923 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159427 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159443 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159453 4923 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159463 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159473 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159482 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159493 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159503 4923 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159514 4923 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159523 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159533 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159544 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159556 4923 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159568 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159545 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159600 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159611 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159581 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159688 4923 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159705 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159724 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159743 4923 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159757 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159774 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159799 4923 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159812 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159825 4923 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159845 4923 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159861 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159876 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.159890 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.159966 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.159986 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.159998 4923 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.163198 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.166848 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.167157 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.167260 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-11 13:24:10.667234492 +0000 UTC m=+98.349136196 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.167246 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.167335 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.167490 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.167629 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.167728 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.167844 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.168465 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.169261 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.169635 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.171140 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.172141 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.178606 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.180715 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.185063 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.193690 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.196121 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.196158 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.196169 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.196187 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.196198 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:10Z","lastTransitionTime":"2026-03-11T13:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.261168 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.261221 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.261470 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.261509 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.261519 4923 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.261530 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.261541 4923 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.261551 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.261561 4923 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.261569 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.261579 4923 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.261589 4923 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.261569 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.261598 4923 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262104 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262126 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262144 4923 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262166 4923 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.261641 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262185 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262386 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262417 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262443 4923 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262464 4923 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262484 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262505 4923 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262527 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262547 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262566 4923 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262586 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262604 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262622 4923 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262641 4923 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262659 4923 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262681 4923 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262701 4923 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262720 4923 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262738 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.262756 4923 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.299500 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.299560 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.299572 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.299595 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.299608 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:10Z","lastTransitionTime":"2026-03-11T13:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.319021 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.325463 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.333511 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.403081 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.403469 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.403483 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.403987 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.404023 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:10Z","lastTransitionTime":"2026-03-11T13:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.463302 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"1d420767a7c5bc4169e1e067b039ec5b88a71a0b3809ca2e7aa57b3199698b54"} Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.465703 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"429cbc6db403c8f599898237353bcb6fc393c120c133bff9483c647cb79de5ed"} Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.467067 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"48b30e461c5f4a21c9bfc2091e3104f8c02b22c1e780f3ed37d87ec17f91d624"} Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.507045 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.507124 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.507139 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.507705 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.507726 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:10Z","lastTransitionTime":"2026-03-11T13:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.610192 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.610255 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.610267 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.610288 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.610301 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:10Z","lastTransitionTime":"2026-03-11T13:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.666214 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.666303 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.666332 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.666368 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.666512 4923 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.666588 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-11 13:24:11.666568867 +0000 UTC m=+99.348470581 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.666595 4923 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.666665 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:11.666611188 +0000 UTC m=+99.348512942 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.666725 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-11 13:24:11.6667045 +0000 UTC m=+99.348606514 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.666769 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.666831 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.666848 4923 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.666927 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-11 13:24:11.666890774 +0000 UTC m=+99.348792488 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.712863 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.712907 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.712917 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.712936 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.712949 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:10Z","lastTransitionTime":"2026-03-11T13:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.766857 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.767093 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.767126 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.767143 4923 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 11 13:24:10 crc kubenswrapper[4923]: E0311 13:24:10.767220 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-11 13:24:11.767194885 +0000 UTC m=+99.449096599 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.815761 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.815811 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.815822 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.815842 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.815857 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:10Z","lastTransitionTime":"2026-03-11T13:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.919010 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.919068 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.919083 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.919111 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:10 crc kubenswrapper[4923]: I0311 13:24:10.919127 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:10Z","lastTransitionTime":"2026-03-11T13:24:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.022516 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.022567 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.022581 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.022601 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.022616 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:11Z","lastTransitionTime":"2026-03-11T13:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.025283 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.025961 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.027698 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.028688 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.030034 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.030857 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.031669 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.032882 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.033670 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.035041 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.035764 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.037188 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.038119 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.038986 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.040241 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.040913 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.042158 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.043049 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.043937 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.045222 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.045836 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.047238 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.047894 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.049248 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.049886 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.051050 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.052858 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.053560 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.054845 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.055496 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.056704 4923 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.056871 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.059174 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.060471 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.061043 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.063283 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.064149 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.065486 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.066383 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.067780 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.068446 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.070159 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.071303 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.073098 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.074376 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.075627 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.076404 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.077910 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.078746 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.079886 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.080598 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.081819 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.082826 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.083533 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.126446 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.126531 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.126555 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.126585 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.126619 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:11Z","lastTransitionTime":"2026-03-11T13:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.194766 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-dmctp"] Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.195143 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-dmctp" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.197998 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.198310 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.198629 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.213054 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.228895 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.228951 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.228970 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.228994 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.229012 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:11Z","lastTransitionTime":"2026-03-11T13:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.257829 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.272617 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/5b0a6b95-e11a-45a1-acb6-34228ee28218-hosts-file\") pod \"node-resolver-dmctp\" (UID: \"5b0a6b95-e11a-45a1-acb6-34228ee28218\") " pod="openshift-dns/node-resolver-dmctp" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.272770 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hshp5\" (UniqueName: \"kubernetes.io/projected/5b0a6b95-e11a-45a1-acb6-34228ee28218-kube-api-access-hshp5\") pod \"node-resolver-dmctp\" (UID: \"5b0a6b95-e11a-45a1-acb6-34228ee28218\") " pod="openshift-dns/node-resolver-dmctp" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.281972 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.300438 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.313706 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dmctp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b0a6b95-e11a-45a1-acb6-34228ee28218\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hshp5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-11T13:24:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dmctp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.331402 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.331442 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.331454 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.331473 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.331486 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:11Z","lastTransitionTime":"2026-03-11T13:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.340084 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.360303 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.374188 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/5b0a6b95-e11a-45a1-acb6-34228ee28218-hosts-file\") pod \"node-resolver-dmctp\" (UID: \"5b0a6b95-e11a-45a1-acb6-34228ee28218\") " pod="openshift-dns/node-resolver-dmctp" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.374290 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hshp5\" (UniqueName: \"kubernetes.io/projected/5b0a6b95-e11a-45a1-acb6-34228ee28218-kube-api-access-hshp5\") pod \"node-resolver-dmctp\" (UID: \"5b0a6b95-e11a-45a1-acb6-34228ee28218\") " pod="openshift-dns/node-resolver-dmctp" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.374708 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/5b0a6b95-e11a-45a1-acb6-34228ee28218-hosts-file\") pod \"node-resolver-dmctp\" (UID: \"5b0a6b95-e11a-45a1-acb6-34228ee28218\") " pod="openshift-dns/node-resolver-dmctp" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.399960 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hshp5\" (UniqueName: \"kubernetes.io/projected/5b0a6b95-e11a-45a1-acb6-34228ee28218-kube-api-access-hshp5\") pod \"node-resolver-dmctp\" (UID: \"5b0a6b95-e11a-45a1-acb6-34228ee28218\") " pod="openshift-dns/node-resolver-dmctp" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.434310 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.434389 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.434406 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.434429 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.434446 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:11Z","lastTransitionTime":"2026-03-11T13:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.471680 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"f3702ba93dfdcaafd55780e2758f74798025167ee922654ecd5442f21a7fd289"} Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.471734 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ab52d6c16bf580573e3d793650f30cfc7430d18dfac0737aba51af3dab86ebc6"} Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.474226 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f4b60f7b311e59c1e16ce2721747d9576d3e393b39e81f9323183c37c3a834f6"} Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.492182 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:09Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.511953 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.513395 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-dmctp" Mar 11 13:24:11 crc kubenswrapper[4923]: W0311 13:24:11.531952 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b0a6b95_e11a_45a1_acb6_34228ee28218.slice/crio-fa12e06f4d92bb328a5f4d04844c239579f9132b0372e4d30eb7739c9e31899a WatchSource:0}: Error finding container fa12e06f4d92bb328a5f4d04844c239579f9132b0372e4d30eb7739c9e31899a: Status 404 returned error can't find the container with id fa12e06f4d92bb328a5f4d04844c239579f9132b0372e4d30eb7739c9e31899a Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.540715 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.540750 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.540763 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.540781 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.540797 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:11Z","lastTransitionTime":"2026-03-11T13:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.541883 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.557717 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.581531 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3702ba93dfdcaafd55780e2758f74798025167ee922654ecd5442f21a7fd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-11T13:24:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab52d6c16bf580573e3d793650f30cfc7430d18dfac0737aba51af3dab86ebc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-11T13:24:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.609257 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.609583 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-fsfnw"] Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.610180 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.610590 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-mx6fg"] Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.611707 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.611849 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-qfw4x"] Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.612222 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.616388 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.616485 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.616720 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.617134 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.618134 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.618539 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.618919 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.618939 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.618986 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.619095 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.619512 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.621527 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.629648 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dmctp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b0a6b95-e11a-45a1-acb6-34228ee28218\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hshp5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-11T13:24:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dmctp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.644654 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.644695 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.644712 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.644735 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.644752 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:11Z","lastTransitionTime":"2026-03-11T13:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.651613 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3702ba93dfdcaafd55780e2758f74798025167ee922654ecd5442f21a7fd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-11T13:24:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab52d6c16bf580573e3d793650f30cfc7430d18dfac0737aba51af3dab86ebc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-11T13:24:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.667479 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dmctp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b0a6b95-e11a-45a1-acb6-34228ee28218\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hshp5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-11T13:24:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dmctp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.677871 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678016 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-system-cni-dir\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: E0311 13:24:11.678078 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:13.678039525 +0000 UTC m=+101.359941289 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678182 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g4pl\" (UniqueName: \"kubernetes.io/projected/e2bd8524-0fce-4b0d-bffb-0013dac2b65f-kube-api-access-5g4pl\") pod \"multus-additional-cni-plugins-mx6fg\" (UID: \"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\") " pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678267 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e2bd8524-0fce-4b0d-bffb-0013dac2b65f-system-cni-dir\") pod \"multus-additional-cni-plugins-mx6fg\" (UID: \"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\") " pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678308 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-host-var-lib-cni-bin\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678388 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgf4z\" (UniqueName: \"kubernetes.io/projected/356de301-ded3-4112-8406-1a28e44c2082-kube-api-access-zgf4z\") pod \"machine-config-daemon-fsfnw\" (UID: \"356de301-ded3-4112-8406-1a28e44c2082\") " pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678428 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-multus-socket-dir-parent\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678471 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678506 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/534d6af5-be40-40a6-accb-8b8b11cbf774-cni-binary-copy\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678538 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-host-run-k8s-cni-cncf-io\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678575 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678613 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e2bd8524-0fce-4b0d-bffb-0013dac2b65f-cni-binary-copy\") pod \"multus-additional-cni-plugins-mx6fg\" (UID: \"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\") " pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678647 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-hostroot\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678680 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e2bd8524-0fce-4b0d-bffb-0013dac2b65f-cnibin\") pod \"multus-additional-cni-plugins-mx6fg\" (UID: \"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\") " pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678707 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-cnibin\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678734 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-os-release\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678761 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-host-run-netns\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678790 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-host-run-multus-certs\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678822 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e2bd8524-0fce-4b0d-bffb-0013dac2b65f-os-release\") pod \"multus-additional-cni-plugins-mx6fg\" (UID: \"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\") " pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678868 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e2bd8524-0fce-4b0d-bffb-0013dac2b65f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-mx6fg\" (UID: \"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\") " pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678905 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678936 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-host-var-lib-cni-multus\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678968 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-multus-conf-dir\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.678998 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/356de301-ded3-4112-8406-1a28e44c2082-proxy-tls\") pod \"machine-config-daemon-fsfnw\" (UID: \"356de301-ded3-4112-8406-1a28e44c2082\") " pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.679028 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-host-var-lib-kubelet\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.679055 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/534d6af5-be40-40a6-accb-8b8b11cbf774-multus-daemon-config\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.679088 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/356de301-ded3-4112-8406-1a28e44c2082-mcd-auth-proxy-config\") pod \"machine-config-daemon-fsfnw\" (UID: \"356de301-ded3-4112-8406-1a28e44c2082\") " pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.679124 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e2bd8524-0fce-4b0d-bffb-0013dac2b65f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-mx6fg\" (UID: \"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\") " pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.679153 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/356de301-ded3-4112-8406-1a28e44c2082-rootfs\") pod \"machine-config-daemon-fsfnw\" (UID: \"356de301-ded3-4112-8406-1a28e44c2082\") " pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.679180 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-etc-kubernetes\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.679213 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-multus-cni-dir\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.679242 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lms6f\" (UniqueName: \"kubernetes.io/projected/534d6af5-be40-40a6-accb-8b8b11cbf774-kube-api-access-lms6f\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: E0311 13:24:11.679474 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 11 13:24:11 crc kubenswrapper[4923]: E0311 13:24:11.679501 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 11 13:24:11 crc kubenswrapper[4923]: E0311 13:24:11.679522 4923 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 11 13:24:11 crc kubenswrapper[4923]: E0311 13:24:11.679584 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-11 13:24:13.679561683 +0000 UTC m=+101.361463437 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 11 13:24:11 crc kubenswrapper[4923]: E0311 13:24:11.679693 4923 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 11 13:24:11 crc kubenswrapper[4923]: E0311 13:24:11.679931 4923 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 11 13:24:11 crc kubenswrapper[4923]: E0311 13:24:11.680026 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-11 13:24:13.679986744 +0000 UTC m=+101.361888578 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 11 13:24:11 crc kubenswrapper[4923]: E0311 13:24:11.681853 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-11 13:24:13.680034895 +0000 UTC m=+101.361936609 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.689832 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mx6fg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g4pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g4pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g4pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g4pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g4pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g4pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g4pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-11T13:24:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mx6fg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.706623 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.722539 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"356de301-ded3-4112-8406-1a28e44c2082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-11T13:24:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fsfnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.739502 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b60f7b311e59c1e16ce2721747d9576d3e393b39e81f9323183c37c3a834f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-11T13:24:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.749660 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.749737 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.749752 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.749776 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.749788 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:11Z","lastTransitionTime":"2026-03-11T13:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.755151 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.773051 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.781155 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.781214 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e2bd8524-0fce-4b0d-bffb-0013dac2b65f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-mx6fg\" (UID: \"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\") " pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.781246 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-host-var-lib-cni-multus\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.781313 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-multus-conf-dir\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.781372 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/356de301-ded3-4112-8406-1a28e44c2082-proxy-tls\") pod \"machine-config-daemon-fsfnw\" (UID: \"356de301-ded3-4112-8406-1a28e44c2082\") " pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.781401 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-host-var-lib-kubelet\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.781430 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/534d6af5-be40-40a6-accb-8b8b11cbf774-multus-daemon-config\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.781461 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e2bd8524-0fce-4b0d-bffb-0013dac2b65f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-mx6fg\" (UID: \"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\") " pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.781488 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/356de301-ded3-4112-8406-1a28e44c2082-mcd-auth-proxy-config\") pod \"machine-config-daemon-fsfnw\" (UID: \"356de301-ded3-4112-8406-1a28e44c2082\") " pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.781512 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/356de301-ded3-4112-8406-1a28e44c2082-rootfs\") pod \"machine-config-daemon-fsfnw\" (UID: \"356de301-ded3-4112-8406-1a28e44c2082\") " pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.781542 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-etc-kubernetes\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.782044 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-multus-cni-dir\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.782102 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lms6f\" (UniqueName: \"kubernetes.io/projected/534d6af5-be40-40a6-accb-8b8b11cbf774-kube-api-access-lms6f\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.782147 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g4pl\" (UniqueName: \"kubernetes.io/projected/e2bd8524-0fce-4b0d-bffb-0013dac2b65f-kube-api-access-5g4pl\") pod \"multus-additional-cni-plugins-mx6fg\" (UID: \"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\") " pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.782142 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-multus-conf-dir\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.782176 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-system-cni-dir\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.782171 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-host-var-lib-cni-multus\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.782136 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/356de301-ded3-4112-8406-1a28e44c2082-rootfs\") pod \"machine-config-daemon-fsfnw\" (UID: \"356de301-ded3-4112-8406-1a28e44c2082\") " pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.782213 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e2bd8524-0fce-4b0d-bffb-0013dac2b65f-system-cni-dir\") pod \"multus-additional-cni-plugins-mx6fg\" (UID: \"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\") " pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.782284 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-host-var-lib-cni-bin\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.782329 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgf4z\" (UniqueName: \"kubernetes.io/projected/356de301-ded3-4112-8406-1a28e44c2082-kube-api-access-zgf4z\") pod \"machine-config-daemon-fsfnw\" (UID: \"356de301-ded3-4112-8406-1a28e44c2082\") " pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" Mar 11 13:24:11 crc kubenswrapper[4923]: E0311 13:24:11.782330 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.782378 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-multus-socket-dir-parent\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: E0311 13:24:11.782402 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 11 13:24:11 crc kubenswrapper[4923]: E0311 13:24:11.782433 4923 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 11 13:24:11 crc kubenswrapper[4923]: E0311 13:24:11.782548 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-11 13:24:13.782512389 +0000 UTC m=+101.464414133 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.782589 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/e2bd8524-0fce-4b0d-bffb-0013dac2b65f-system-cni-dir\") pod \"multus-additional-cni-plugins-mx6fg\" (UID: \"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\") " pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.782649 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-etc-kubernetes\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.782674 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/e2bd8524-0fce-4b0d-bffb-0013dac2b65f-tuning-conf-dir\") pod \"multus-additional-cni-plugins-mx6fg\" (UID: \"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\") " pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.782716 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-multus-cni-dir\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.782825 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-host-var-lib-kubelet\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.782431 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/534d6af5-be40-40a6-accb-8b8b11cbf774-cni-binary-copy\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.782932 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-host-run-k8s-cni-cncf-io\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.783102 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/356de301-ded3-4112-8406-1a28e44c2082-mcd-auth-proxy-config\") pod \"machine-config-daemon-fsfnw\" (UID: \"356de301-ded3-4112-8406-1a28e44c2082\") " pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.783118 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-host-var-lib-cni-bin\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.783133 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/e2bd8524-0fce-4b0d-bffb-0013dac2b65f-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-mx6fg\" (UID: \"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\") " pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.783187 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-host-run-k8s-cni-cncf-io\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.783197 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/534d6af5-be40-40a6-accb-8b8b11cbf774-cni-binary-copy\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.783226 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-multus-socket-dir-parent\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.783322 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e2bd8524-0fce-4b0d-bffb-0013dac2b65f-cni-binary-copy\") pod \"multus-additional-cni-plugins-mx6fg\" (UID: \"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\") " pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.783431 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-hostroot\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.783500 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e2bd8524-0fce-4b0d-bffb-0013dac2b65f-cnibin\") pod \"multus-additional-cni-plugins-mx6fg\" (UID: \"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\") " pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.783540 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e2bd8524-0fce-4b0d-bffb-0013dac2b65f-os-release\") pod \"multus-additional-cni-plugins-mx6fg\" (UID: \"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\") " pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.783571 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-cnibin\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.783573 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-hostroot\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.783600 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-os-release\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.783627 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-host-run-netns\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.783651 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-host-run-multus-certs\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.783672 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-cnibin\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.783689 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/e2bd8524-0fce-4b0d-bffb-0013dac2b65f-os-release\") pod \"multus-additional-cni-plugins-mx6fg\" (UID: \"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\") " pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.783731 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/e2bd8524-0fce-4b0d-bffb-0013dac2b65f-cnibin\") pod \"multus-additional-cni-plugins-mx6fg\" (UID: \"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\") " pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.783748 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-host-run-netns\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.783748 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-os-release\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.783801 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-host-run-multus-certs\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.782963 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/534d6af5-be40-40a6-accb-8b8b11cbf774-system-cni-dir\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.784220 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/e2bd8524-0fce-4b0d-bffb-0013dac2b65f-cni-binary-copy\") pod \"multus-additional-cni-plugins-mx6fg\" (UID: \"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\") " pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.784209 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/534d6af5-be40-40a6-accb-8b8b11cbf774-multus-daemon-config\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.787843 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/356de301-ded3-4112-8406-1a28e44c2082-proxy-tls\") pod \"machine-config-daemon-fsfnw\" (UID: \"356de301-ded3-4112-8406-1a28e44c2082\") " pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.796679 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.800268 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgf4z\" (UniqueName: \"kubernetes.io/projected/356de301-ded3-4112-8406-1a28e44c2082-kube-api-access-zgf4z\") pod \"machine-config-daemon-fsfnw\" (UID: \"356de301-ded3-4112-8406-1a28e44c2082\") " pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.803191 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g4pl\" (UniqueName: \"kubernetes.io/projected/e2bd8524-0fce-4b0d-bffb-0013dac2b65f-kube-api-access-5g4pl\") pod \"multus-additional-cni-plugins-mx6fg\" (UID: \"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\") " pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.806589 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lms6f\" (UniqueName: \"kubernetes.io/projected/534d6af5-be40-40a6-accb-8b8b11cbf774-kube-api-access-lms6f\") pod \"multus-qfw4x\" (UID: \"534d6af5-be40-40a6-accb-8b8b11cbf774\") " pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.814454 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfw4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"534d6af5-be40-40a6-accb-8b8b11cbf774\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lms6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-11T13:24:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfw4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:11Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.852739 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.852787 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.852802 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.852820 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.852832 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:11Z","lastTransitionTime":"2026-03-11T13:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.931211 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.943862 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-mx6fg" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.951493 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-qfw4x" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.955399 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.955555 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.955656 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.955772 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:11 crc kubenswrapper[4923]: I0311 13:24:11.955875 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:11Z","lastTransitionTime":"2026-03-11T13:24:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.005980 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-552rb"] Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.007222 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.010099 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.012977 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.013299 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.013904 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.014665 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.014857 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.015189 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.021149 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:12 crc kubenswrapper[4923]: E0311 13:24:12.021256 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.021320 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.021375 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 11 13:24:12 crc kubenswrapper[4923]: E0311 13:24:12.021520 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 11 13:24:12 crc kubenswrapper[4923]: E0311 13:24:12.021644 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.034335 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfw4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"534d6af5-be40-40a6-accb-8b8b11cbf774\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lms6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-11T13:24:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfw4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:12Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.053459 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.055951 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:12Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.058058 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.058158 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.058215 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.058278 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.058355 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:12Z","lastTransitionTime":"2026-03-11T13:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.078892 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f3702ba93dfdcaafd55780e2758f74798025167ee922654ecd5442f21a7fd289\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-11T13:24:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab52d6c16bf580573e3d793650f30cfc7430d18dfac0737aba51af3dab86ebc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-11T13:24:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:12Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:12 crc kubenswrapper[4923]: W0311 13:24:12.080544 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod356de301_ded3_4112_8406_1a28e44c2082.slice/crio-425dbda924b4037a5434f0daf7d540c1e582969f0820906e203c44a683474d70 WatchSource:0}: Error finding container 425dbda924b4037a5434f0daf7d540c1e582969f0820906e203c44a683474d70: Status 404 returned error can't find the container with id 425dbda924b4037a5434f0daf7d540c1e582969f0820906e203c44a683474d70 Mar 11 13:24:12 crc kubenswrapper[4923]: W0311 13:24:12.082149 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2bd8524_0fce_4b0d_bffb_0013dac2b65f.slice/crio-b5b1cc4a20d88b598d9f235ca9e26c3ec19577c4980fec6782a06ea84c75784d WatchSource:0}: Error finding container b5b1cc4a20d88b598d9f235ca9e26c3ec19577c4980fec6782a06ea84c75784d: Status 404 returned error can't find the container with id b5b1cc4a20d88b598d9f235ca9e26c3ec19577c4980fec6782a06ea84c75784d Mar 11 13:24:12 crc kubenswrapper[4923]: W0311 13:24:12.084294 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod534d6af5_be40_40a6_accb_8b8b11cbf774.slice/crio-5a7f61785480d84b2c1b490d19c6dd1f7d6f0145972e95a1a0567fe402d50576 WatchSource:0}: Error finding container 5a7f61785480d84b2c1b490d19c6dd1f7d6f0145972e95a1a0567fe402d50576: Status 404 returned error can't find the container with id 5a7f61785480d84b2c1b490d19c6dd1f7d6f0145972e95a1a0567fe402d50576 Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.085525 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/df02f0a6-797a-4c04-a529-f0881241fd94-ovn-node-metrics-cert\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.085555 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-run-ovn\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.085574 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.085591 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-log-socket\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.085611 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-kubelet\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.085667 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-run-ovn-kubernetes\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.085712 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22njz\" (UniqueName: \"kubernetes.io/projected/df02f0a6-797a-4c04-a529-f0881241fd94-kube-api-access-22njz\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.086005 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-etc-openvswitch\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.086033 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-systemd-units\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.086053 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/df02f0a6-797a-4c04-a529-f0881241fd94-env-overrides\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.086082 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-cni-bin\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.086099 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/df02f0a6-797a-4c04-a529-f0881241fd94-ovnkube-script-lib\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.086167 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-slash\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.086273 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-run-openvswitch\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.086332 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-run-netns\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.086386 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-var-lib-openvswitch\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.086444 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-cni-netd\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.086488 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-run-systemd\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.086542 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-node-log\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.086565 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/df02f0a6-797a-4c04-a529-f0881241fd94-ovnkube-config\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.094353 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-dmctp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b0a6b95-e11a-45a1-acb6-34228ee28218\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hshp5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-11T13:24:11Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-dmctp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:12Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.114635 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mx6fg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e2bd8524-0fce-4b0d-bffb-0013dac2b65f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g4pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g4pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g4pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g4pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g4pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g4pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5g4pl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-11T13:24:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mx6fg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:12Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.131360 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b60f7b311e59c1e16ce2721747d9576d3e393b39e81f9323183c37c3a834f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-11T13:24:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:12Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.160753 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.160824 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.160836 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.160858 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.160870 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:12Z","lastTransitionTime":"2026-03-11T13:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.169867 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:12Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.185288 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"356de301-ded3-4112-8406-1a28e44c2082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-11T13:24:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fsfnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:12Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.187869 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-node-log\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.187932 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/df02f0a6-797a-4c04-a529-f0881241fd94-ovnkube-config\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.187965 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/df02f0a6-797a-4c04-a529-f0881241fd94-ovn-node-metrics-cert\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.187987 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-run-ovn\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188014 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188037 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-log-socket\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188042 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-node-log\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188108 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-kubelet\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188056 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-kubelet\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188146 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188160 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-run-ovn-kubernetes\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188188 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-run-ovn\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188188 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22njz\" (UniqueName: \"kubernetes.io/projected/df02f0a6-797a-4c04-a529-f0881241fd94-kube-api-access-22njz\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188253 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-etc-openvswitch\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188290 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-systemd-units\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188315 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/df02f0a6-797a-4c04-a529-f0881241fd94-env-overrides\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188380 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-cni-bin\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188407 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/df02f0a6-797a-4c04-a529-f0881241fd94-ovnkube-script-lib\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188452 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-slash\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188476 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-run-openvswitch\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188515 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-run-netns\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188537 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-var-lib-openvswitch\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188564 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-cni-netd\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188596 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-run-systemd\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188615 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-log-socket\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188676 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-var-lib-openvswitch\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188704 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-run-openvswitch\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188754 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-run-netns\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188779 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-slash\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188800 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-etc-openvswitch\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188822 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-systemd-units\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188854 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-run-systemd\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188883 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-cni-netd\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188906 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-run-ovn-kubernetes\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.188949 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-cni-bin\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.189238 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/df02f0a6-797a-4c04-a529-f0881241fd94-ovnkube-config\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.190588 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/df02f0a6-797a-4c04-a529-f0881241fd94-env-overrides\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.193686 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/df02f0a6-797a-4c04-a529-f0881241fd94-ovnkube-script-lib\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.196373 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/df02f0a6-797a-4c04-a529-f0881241fd94-ovn-node-metrics-cert\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.205061 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22njz\" (UniqueName: \"kubernetes.io/projected/df02f0a6-797a-4c04-a529-f0881241fd94-kube-api-access-22njz\") pod \"ovnkube-node-552rb\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.208449 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df02f0a6-797a-4c04-a529-f0881241fd94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:12Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:12Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22njz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22njz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22njz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22njz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22njz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22njz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22njz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22njz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22njz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-11T13:24:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-552rb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:12Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.221165 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:12Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.233358 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:12Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.264566 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.264605 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.264617 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.264633 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.264645 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:12Z","lastTransitionTime":"2026-03-11T13:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.328072 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.368393 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.368566 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.368647 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.368722 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.368780 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:12Z","lastTransitionTime":"2026-03-11T13:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.471567 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.471601 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.471608 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.471624 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.471634 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:12Z","lastTransitionTime":"2026-03-11T13:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.478398 4923 generic.go:334] "Generic (PLEG): container finished" podID="df02f0a6-797a-4c04-a529-f0881241fd94" containerID="c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739" exitCode=0 Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.478443 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" event={"ID":"df02f0a6-797a-4c04-a529-f0881241fd94","Type":"ContainerDied","Data":"c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739"} Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.478506 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" event={"ID":"df02f0a6-797a-4c04-a529-f0881241fd94","Type":"ContainerStarted","Data":"417cfa13eee32fab3ba6faf38c6585235ab1ae14743d4aa89ac68014dedfb0aa"} Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.481269 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" event={"ID":"356de301-ded3-4112-8406-1a28e44c2082","Type":"ContainerStarted","Data":"9ec5b976ec23960ccba857a6be05968b9aa3d76c1cba14b8c9e6598ca64ebc85"} Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.481305 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" event={"ID":"356de301-ded3-4112-8406-1a28e44c2082","Type":"ContainerStarted","Data":"e04ff4641c7ac0b2e854199edde21a639f88c1a928f8af341f17e144a3a65e5d"} Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.481320 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" event={"ID":"356de301-ded3-4112-8406-1a28e44c2082","Type":"ContainerStarted","Data":"425dbda924b4037a5434f0daf7d540c1e582969f0820906e203c44a683474d70"} Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.483996 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-dmctp" event={"ID":"5b0a6b95-e11a-45a1-acb6-34228ee28218","Type":"ContainerStarted","Data":"502c943da27f3e3c2b77478cf5626049556634676d9412354ce23b2affaa762d"} Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.484040 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-dmctp" event={"ID":"5b0a6b95-e11a-45a1-acb6-34228ee28218","Type":"ContainerStarted","Data":"fa12e06f4d92bb328a5f4d04844c239579f9132b0372e4d30eb7739c9e31899a"} Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.486274 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qfw4x" event={"ID":"534d6af5-be40-40a6-accb-8b8b11cbf774","Type":"ContainerStarted","Data":"85cc2cd095a7a54377400f5d541dc109c15e09e067012495defd5aa6e2ec9ee3"} Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.486463 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qfw4x" event={"ID":"534d6af5-be40-40a6-accb-8b8b11cbf774","Type":"ContainerStarted","Data":"5a7f61785480d84b2c1b490d19c6dd1f7d6f0145972e95a1a0567fe402d50576"} Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.488622 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mx6fg" event={"ID":"e2bd8524-0fce-4b0d-bffb-0013dac2b65f","Type":"ContainerStarted","Data":"d117c35c5700d56f6af0e439a53692697fc2dac450c9c804069184bbe6ac1d8c"} Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.488739 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mx6fg" event={"ID":"e2bd8524-0fce-4b0d-bffb-0013dac2b65f","Type":"ContainerStarted","Data":"b5b1cc4a20d88b598d9f235ca9e26c3ec19577c4980fec6782a06ea84c75784d"} Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.495767 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"356de301-ded3-4112-8406-1a28e44c2082\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zgf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-11T13:24:11Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-fsfnw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:12Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.516026 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"df02f0a6-797a-4c04-a529-f0881241fd94\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:12Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22njz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22njz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22njz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22njz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22njz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22njz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22njz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22njz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-11T13:24:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-11T13:24:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-22njz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-11T13:24:12Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-552rb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:12Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.533252 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f4b60f7b311e59c1e16ce2721747d9576d3e393b39e81f9323183c37c3a834f6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-11T13:24:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:12Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.548201 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:12Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.560413 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:12Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.572682 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:12Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.574477 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.574516 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.574526 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.574542 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.574554 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:12Z","lastTransitionTime":"2026-03-11T13:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.593579 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qfw4x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"534d6af5-be40-40a6-accb-8b8b11cbf774\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-11T13:24:11Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lms6f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-11T13:24:11Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qfw4x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-11T13:24:12Z is after 2025-08-24T17:21:41Z" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.676328 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.676381 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.676393 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.676409 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.676420 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:12Z","lastTransitionTime":"2026-03-11T13:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.687755 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=0.687729 podStartE2EDuration="687.729ms" podCreationTimestamp="2026-03-11 13:24:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:12.68615612 +0000 UTC m=+100.368057864" watchObservedRunningTime="2026-03-11 13:24:12.687729 +0000 UTC m=+100.369630734" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.737694 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-dmctp" podStartSLOduration=47.73767226 podStartE2EDuration="47.73767226s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:12.713185442 +0000 UTC m=+100.395087156" watchObservedRunningTime="2026-03-11 13:24:12.73767226 +0000 UTC m=+100.419573984" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.779283 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.779318 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.779328 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.779355 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.779400 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:12Z","lastTransitionTime":"2026-03-11T13:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.796853 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podStartSLOduration=47.796830708 podStartE2EDuration="47.796830708s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:12.77556924 +0000 UTC m=+100.457470964" watchObservedRunningTime="2026-03-11 13:24:12.796830708 +0000 UTC m=+100.478732432" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.881953 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.881996 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.882012 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.882030 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.882042 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:12Z","lastTransitionTime":"2026-03-11T13:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.914716 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-qfw4x" podStartSLOduration=47.914695294 podStartE2EDuration="47.914695294s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:12.838727128 +0000 UTC m=+100.520628852" watchObservedRunningTime="2026-03-11 13:24:12.914695294 +0000 UTC m=+100.596597008" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.915317 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-rlsnm"] Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.915792 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-rlsnm" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.917825 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.918235 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.918407 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.918760 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.984922 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.984961 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.984973 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.984991 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.985006 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:12Z","lastTransitionTime":"2026-03-11T13:24:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.996690 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/51eb36cc-7a58-4041-9419-9b71fc58a396-host\") pod \"node-ca-rlsnm\" (UID: \"51eb36cc-7a58-4041-9419-9b71fc58a396\") " pod="openshift-image-registry/node-ca-rlsnm" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.996749 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27nb2\" (UniqueName: \"kubernetes.io/projected/51eb36cc-7a58-4041-9419-9b71fc58a396-kube-api-access-27nb2\") pod \"node-ca-rlsnm\" (UID: \"51eb36cc-7a58-4041-9419-9b71fc58a396\") " pod="openshift-image-registry/node-ca-rlsnm" Mar 11 13:24:12 crc kubenswrapper[4923]: I0311 13:24:12.996783 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/51eb36cc-7a58-4041-9419-9b71fc58a396-serviceca\") pod \"node-ca-rlsnm\" (UID: \"51eb36cc-7a58-4041-9419-9b71fc58a396\") " pod="openshift-image-registry/node-ca-rlsnm" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.083944 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hxtts"] Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.084415 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hxtts" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.086492 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.086664 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.086696 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.086708 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.086724 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.086737 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:13Z","lastTransitionTime":"2026-03-11T13:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.087169 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.097267 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27nb2\" (UniqueName: \"kubernetes.io/projected/51eb36cc-7a58-4041-9419-9b71fc58a396-kube-api-access-27nb2\") pod \"node-ca-rlsnm\" (UID: \"51eb36cc-7a58-4041-9419-9b71fc58a396\") " pod="openshift-image-registry/node-ca-rlsnm" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.097323 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/51eb36cc-7a58-4041-9419-9b71fc58a396-serviceca\") pod \"node-ca-rlsnm\" (UID: \"51eb36cc-7a58-4041-9419-9b71fc58a396\") " pod="openshift-image-registry/node-ca-rlsnm" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.097378 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/51eb36cc-7a58-4041-9419-9b71fc58a396-host\") pod \"node-ca-rlsnm\" (UID: \"51eb36cc-7a58-4041-9419-9b71fc58a396\") " pod="openshift-image-registry/node-ca-rlsnm" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.097451 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/51eb36cc-7a58-4041-9419-9b71fc58a396-host\") pod \"node-ca-rlsnm\" (UID: \"51eb36cc-7a58-4041-9419-9b71fc58a396\") " pod="openshift-image-registry/node-ca-rlsnm" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.098407 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/51eb36cc-7a58-4041-9419-9b71fc58a396-serviceca\") pod \"node-ca-rlsnm\" (UID: \"51eb36cc-7a58-4041-9419-9b71fc58a396\") " pod="openshift-image-registry/node-ca-rlsnm" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.099578 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-4nx54"] Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.100068 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4nx54" Mar 11 13:24:13 crc kubenswrapper[4923]: E0311 13:24:13.100140 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4nx54" podUID="ae7e45f7-8c08-435b-990a-5149da33010a" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.119749 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27nb2\" (UniqueName: \"kubernetes.io/projected/51eb36cc-7a58-4041-9419-9b71fc58a396-kube-api-access-27nb2\") pod \"node-ca-rlsnm\" (UID: \"51eb36cc-7a58-4041-9419-9b71fc58a396\") " pod="openshift-image-registry/node-ca-rlsnm" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.189718 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.189780 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.189790 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.189805 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.189815 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:13Z","lastTransitionTime":"2026-03-11T13:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.198325 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mx8kb\" (UniqueName: \"kubernetes.io/projected/eabd008f-6584-4f8e-9f4c-3fa587891713-kube-api-access-mx8kb\") pod \"ovnkube-control-plane-749d76644c-hxtts\" (UID: \"eabd008f-6584-4f8e-9f4c-3fa587891713\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hxtts" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.198388 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eabd008f-6584-4f8e-9f4c-3fa587891713-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-hxtts\" (UID: \"eabd008f-6584-4f8e-9f4c-3fa587891713\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hxtts" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.198406 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eabd008f-6584-4f8e-9f4c-3fa587891713-env-overrides\") pod \"ovnkube-control-plane-749d76644c-hxtts\" (UID: \"eabd008f-6584-4f8e-9f4c-3fa587891713\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hxtts" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.198426 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bljh5\" (UniqueName: \"kubernetes.io/projected/ae7e45f7-8c08-435b-990a-5149da33010a-kube-api-access-bljh5\") pod \"network-metrics-daemon-4nx54\" (UID: \"ae7e45f7-8c08-435b-990a-5149da33010a\") " pod="openshift-multus/network-metrics-daemon-4nx54" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.198444 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eabd008f-6584-4f8e-9f4c-3fa587891713-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-hxtts\" (UID: \"eabd008f-6584-4f8e-9f4c-3fa587891713\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hxtts" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.198572 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ae7e45f7-8c08-435b-990a-5149da33010a-metrics-certs\") pod \"network-metrics-daemon-4nx54\" (UID: \"ae7e45f7-8c08-435b-990a-5149da33010a\") " pod="openshift-multus/network-metrics-daemon-4nx54" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.254323 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-rlsnm" Mar 11 13:24:13 crc kubenswrapper[4923]: W0311 13:24:13.277229 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51eb36cc_7a58_4041_9419_9b71fc58a396.slice/crio-991c57a22121c7e18cd44fee2662a9ad547736fe79dedee0fb7c297b8bc8c2c6 WatchSource:0}: Error finding container 991c57a22121c7e18cd44fee2662a9ad547736fe79dedee0fb7c297b8bc8c2c6: Status 404 returned error can't find the container with id 991c57a22121c7e18cd44fee2662a9ad547736fe79dedee0fb7c297b8bc8c2c6 Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.292117 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.292274 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.292381 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.292453 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.292512 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:13Z","lastTransitionTime":"2026-03-11T13:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.299449 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bljh5\" (UniqueName: \"kubernetes.io/projected/ae7e45f7-8c08-435b-990a-5149da33010a-kube-api-access-bljh5\") pod \"network-metrics-daemon-4nx54\" (UID: \"ae7e45f7-8c08-435b-990a-5149da33010a\") " pod="openshift-multus/network-metrics-daemon-4nx54" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.299721 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eabd008f-6584-4f8e-9f4c-3fa587891713-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-hxtts\" (UID: \"eabd008f-6584-4f8e-9f4c-3fa587891713\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hxtts" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.299826 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ae7e45f7-8c08-435b-990a-5149da33010a-metrics-certs\") pod \"network-metrics-daemon-4nx54\" (UID: \"ae7e45f7-8c08-435b-990a-5149da33010a\") " pod="openshift-multus/network-metrics-daemon-4nx54" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.299901 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mx8kb\" (UniqueName: \"kubernetes.io/projected/eabd008f-6584-4f8e-9f4c-3fa587891713-kube-api-access-mx8kb\") pod \"ovnkube-control-plane-749d76644c-hxtts\" (UID: \"eabd008f-6584-4f8e-9f4c-3fa587891713\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hxtts" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.299990 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eabd008f-6584-4f8e-9f4c-3fa587891713-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-hxtts\" (UID: \"eabd008f-6584-4f8e-9f4c-3fa587891713\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hxtts" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.300074 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eabd008f-6584-4f8e-9f4c-3fa587891713-env-overrides\") pod \"ovnkube-control-plane-749d76644c-hxtts\" (UID: \"eabd008f-6584-4f8e-9f4c-3fa587891713\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hxtts" Mar 11 13:24:13 crc kubenswrapper[4923]: E0311 13:24:13.299989 4923 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 11 13:24:13 crc kubenswrapper[4923]: E0311 13:24:13.300241 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ae7e45f7-8c08-435b-990a-5149da33010a-metrics-certs podName:ae7e45f7-8c08-435b-990a-5149da33010a nodeName:}" failed. No retries permitted until 2026-03-11 13:24:13.800226284 +0000 UTC m=+101.482127998 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ae7e45f7-8c08-435b-990a-5149da33010a-metrics-certs") pod "network-metrics-daemon-4nx54" (UID: "ae7e45f7-8c08-435b-990a-5149da33010a") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.300869 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/eabd008f-6584-4f8e-9f4c-3fa587891713-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-hxtts\" (UID: \"eabd008f-6584-4f8e-9f4c-3fa587891713\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hxtts" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.302273 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/eabd008f-6584-4f8e-9f4c-3fa587891713-env-overrides\") pod \"ovnkube-control-plane-749d76644c-hxtts\" (UID: \"eabd008f-6584-4f8e-9f4c-3fa587891713\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hxtts" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.305313 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/eabd008f-6584-4f8e-9f4c-3fa587891713-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-hxtts\" (UID: \"eabd008f-6584-4f8e-9f4c-3fa587891713\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hxtts" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.318747 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bljh5\" (UniqueName: \"kubernetes.io/projected/ae7e45f7-8c08-435b-990a-5149da33010a-kube-api-access-bljh5\") pod \"network-metrics-daemon-4nx54\" (UID: \"ae7e45f7-8c08-435b-990a-5149da33010a\") " pod="openshift-multus/network-metrics-daemon-4nx54" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.319331 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mx8kb\" (UniqueName: \"kubernetes.io/projected/eabd008f-6584-4f8e-9f4c-3fa587891713-kube-api-access-mx8kb\") pod \"ovnkube-control-plane-749d76644c-hxtts\" (UID: \"eabd008f-6584-4f8e-9f4c-3fa587891713\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hxtts" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.395191 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.395227 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.395236 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.395250 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.395260 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:13Z","lastTransitionTime":"2026-03-11T13:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.402605 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hxtts" Mar 11 13:24:13 crc kubenswrapper[4923]: W0311 13:24:13.421451 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeabd008f_6584_4f8e_9f4c_3fa587891713.slice/crio-db2fe5a7d6c6cf9c2b6a029b24e0cfbb406e9f1055d31d1d298de833f0c6c086 WatchSource:0}: Error finding container db2fe5a7d6c6cf9c2b6a029b24e0cfbb406e9f1055d31d1d298de833f0c6c086: Status 404 returned error can't find the container with id db2fe5a7d6c6cf9c2b6a029b24e0cfbb406e9f1055d31d1d298de833f0c6c086 Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.496096 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-rlsnm" event={"ID":"51eb36cc-7a58-4041-9419-9b71fc58a396","Type":"ContainerStarted","Data":"991c57a22121c7e18cd44fee2662a9ad547736fe79dedee0fb7c297b8bc8c2c6"} Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.497556 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.497591 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.497599 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.497612 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.497624 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:13Z","lastTransitionTime":"2026-03-11T13:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.502701 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" event={"ID":"df02f0a6-797a-4c04-a529-f0881241fd94","Type":"ContainerStarted","Data":"8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e"} Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.502763 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" event={"ID":"df02f0a6-797a-4c04-a529-f0881241fd94","Type":"ContainerStarted","Data":"8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0"} Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.502785 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" event={"ID":"df02f0a6-797a-4c04-a529-f0881241fd94","Type":"ContainerStarted","Data":"eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e"} Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.502805 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" event={"ID":"df02f0a6-797a-4c04-a529-f0881241fd94","Type":"ContainerStarted","Data":"a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d"} Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.502823 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" event={"ID":"df02f0a6-797a-4c04-a529-f0881241fd94","Type":"ContainerStarted","Data":"6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595"} Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.504333 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"88bc5542d258659a50787cbba1fc73dcb6a5ee7414cba241f4cedbf62b240bc9"} Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.505622 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hxtts" event={"ID":"eabd008f-6584-4f8e-9f4c-3fa587891713","Type":"ContainerStarted","Data":"db2fe5a7d6c6cf9c2b6a029b24e0cfbb406e9f1055d31d1d298de833f0c6c086"} Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.517043 4923 generic.go:334] "Generic (PLEG): container finished" podID="e2bd8524-0fce-4b0d-bffb-0013dac2b65f" containerID="d117c35c5700d56f6af0e439a53692697fc2dac450c9c804069184bbe6ac1d8c" exitCode=0 Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.517092 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mx6fg" event={"ID":"e2bd8524-0fce-4b0d-bffb-0013dac2b65f","Type":"ContainerDied","Data":"d117c35c5700d56f6af0e439a53692697fc2dac450c9c804069184bbe6ac1d8c"} Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.599740 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.599769 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.599778 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.599789 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.599800 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:13Z","lastTransitionTime":"2026-03-11T13:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.702687 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.702742 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.702761 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.702785 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.702804 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:13Z","lastTransitionTime":"2026-03-11T13:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.702873 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.702982 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.703006 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.703045 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:13 crc kubenswrapper[4923]: E0311 13:24:13.703122 4923 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 11 13:24:13 crc kubenswrapper[4923]: E0311 13:24:13.703163 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-11 13:24:17.703151646 +0000 UTC m=+105.385053360 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 11 13:24:13 crc kubenswrapper[4923]: E0311 13:24:13.703217 4923 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 11 13:24:13 crc kubenswrapper[4923]: E0311 13:24:13.703234 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 11 13:24:13 crc kubenswrapper[4923]: E0311 13:24:13.703252 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 11 13:24:13 crc kubenswrapper[4923]: E0311 13:24:13.703253 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-11 13:24:17.703244188 +0000 UTC m=+105.385145902 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 11 13:24:13 crc kubenswrapper[4923]: E0311 13:24:13.703263 4923 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 11 13:24:13 crc kubenswrapper[4923]: E0311 13:24:13.703278 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:17.703269479 +0000 UTC m=+105.385171193 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:13 crc kubenswrapper[4923]: E0311 13:24:13.703295 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-11 13:24:17.703285859 +0000 UTC m=+105.385187573 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.803719 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ae7e45f7-8c08-435b-990a-5149da33010a-metrics-certs\") pod \"network-metrics-daemon-4nx54\" (UID: \"ae7e45f7-8c08-435b-990a-5149da33010a\") " pod="openshift-multus/network-metrics-daemon-4nx54" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.804203 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 11 13:24:13 crc kubenswrapper[4923]: E0311 13:24:13.804438 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 11 13:24:13 crc kubenswrapper[4923]: E0311 13:24:13.804462 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 11 13:24:13 crc kubenswrapper[4923]: E0311 13:24:13.804477 4923 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 11 13:24:13 crc kubenswrapper[4923]: E0311 13:24:13.804527 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-11 13:24:17.804505812 +0000 UTC m=+105.486407526 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 11 13:24:13 crc kubenswrapper[4923]: E0311 13:24:13.804593 4923 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 11 13:24:13 crc kubenswrapper[4923]: E0311 13:24:13.804623 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ae7e45f7-8c08-435b-990a-5149da33010a-metrics-certs podName:ae7e45f7-8c08-435b-990a-5149da33010a nodeName:}" failed. No retries permitted until 2026-03-11 13:24:14.804611964 +0000 UTC m=+102.486513678 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ae7e45f7-8c08-435b-990a-5149da33010a-metrics-certs") pod "network-metrics-daemon-4nx54" (UID: "ae7e45f7-8c08-435b-990a-5149da33010a") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.807979 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.808006 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.808018 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.808036 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.808049 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:13Z","lastTransitionTime":"2026-03-11T13:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.877636 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.877688 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.877701 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.877721 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.877757 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-11T13:24:13Z","lastTransitionTime":"2026-03-11T13:24:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.925799 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-fjqzd"] Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.926231 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fjqzd" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.929327 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.929560 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.929847 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 11 13:24:13 crc kubenswrapper[4923]: I0311 13:24:13.929968 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.009626 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c2b137ab-90be-40d8-9317-7bf25b183d6b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fjqzd\" (UID: \"c2b137ab-90be-40d8-9317-7bf25b183d6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fjqzd" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.009862 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c2b137ab-90be-40d8-9317-7bf25b183d6b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fjqzd\" (UID: \"c2b137ab-90be-40d8-9317-7bf25b183d6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fjqzd" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.009946 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2b137ab-90be-40d8-9317-7bf25b183d6b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fjqzd\" (UID: \"c2b137ab-90be-40d8-9317-7bf25b183d6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fjqzd" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.010011 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c2b137ab-90be-40d8-9317-7bf25b183d6b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fjqzd\" (UID: \"c2b137ab-90be-40d8-9317-7bf25b183d6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fjqzd" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.010065 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2b137ab-90be-40d8-9317-7bf25b183d6b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fjqzd\" (UID: \"c2b137ab-90be-40d8-9317-7bf25b183d6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fjqzd" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.020260 4923 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.020954 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.020968 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.021008 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 11 13:24:14 crc kubenswrapper[4923]: E0311 13:24:14.021103 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 11 13:24:14 crc kubenswrapper[4923]: E0311 13:24:14.021224 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 11 13:24:14 crc kubenswrapper[4923]: E0311 13:24:14.021311 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.031263 4923 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.111459 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c2b137ab-90be-40d8-9317-7bf25b183d6b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fjqzd\" (UID: \"c2b137ab-90be-40d8-9317-7bf25b183d6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fjqzd" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.111499 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2b137ab-90be-40d8-9317-7bf25b183d6b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fjqzd\" (UID: \"c2b137ab-90be-40d8-9317-7bf25b183d6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fjqzd" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.111544 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c2b137ab-90be-40d8-9317-7bf25b183d6b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fjqzd\" (UID: \"c2b137ab-90be-40d8-9317-7bf25b183d6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fjqzd" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.111580 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c2b137ab-90be-40d8-9317-7bf25b183d6b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fjqzd\" (UID: \"c2b137ab-90be-40d8-9317-7bf25b183d6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fjqzd" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.111597 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2b137ab-90be-40d8-9317-7bf25b183d6b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fjqzd\" (UID: \"c2b137ab-90be-40d8-9317-7bf25b183d6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fjqzd" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.111673 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/c2b137ab-90be-40d8-9317-7bf25b183d6b-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fjqzd\" (UID: \"c2b137ab-90be-40d8-9317-7bf25b183d6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fjqzd" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.111848 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/c2b137ab-90be-40d8-9317-7bf25b183d6b-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fjqzd\" (UID: \"c2b137ab-90be-40d8-9317-7bf25b183d6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fjqzd" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.112690 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/c2b137ab-90be-40d8-9317-7bf25b183d6b-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fjqzd\" (UID: \"c2b137ab-90be-40d8-9317-7bf25b183d6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fjqzd" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.119035 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c2b137ab-90be-40d8-9317-7bf25b183d6b-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fjqzd\" (UID: \"c2b137ab-90be-40d8-9317-7bf25b183d6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fjqzd" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.128492 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c2b137ab-90be-40d8-9317-7bf25b183d6b-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fjqzd\" (UID: \"c2b137ab-90be-40d8-9317-7bf25b183d6b\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fjqzd" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.241669 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fjqzd" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.523730 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hxtts" event={"ID":"eabd008f-6584-4f8e-9f4c-3fa587891713","Type":"ContainerStarted","Data":"af5ab7ac9bf86e409f06e54fe2907fd1b95272b9a1a6144954d5fbcf3b61dd5e"} Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.523796 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hxtts" event={"ID":"eabd008f-6584-4f8e-9f4c-3fa587891713","Type":"ContainerStarted","Data":"cb97a990e8b9c2ec2b4d6c78b2707bd3281d9dc1d7eec6e87eb66047ac7a4d70"} Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.527138 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fjqzd" event={"ID":"c2b137ab-90be-40d8-9317-7bf25b183d6b","Type":"ContainerStarted","Data":"57c1c8c303878e8b152d083c90c048c55ae1ba5668d375019024e5a7065bbcc4"} Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.527187 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fjqzd" event={"ID":"c2b137ab-90be-40d8-9317-7bf25b183d6b","Type":"ContainerStarted","Data":"e3dae241b643782678b88a59313d4c935cb81ff6423b5bcac2a46b024de2c054"} Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.529551 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-rlsnm" event={"ID":"51eb36cc-7a58-4041-9419-9b71fc58a396","Type":"ContainerStarted","Data":"fabe26d25944105c2d8cd48545dad164353144cf74da70da49ad94da03200d41"} Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.536805 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" event={"ID":"df02f0a6-797a-4c04-a529-f0881241fd94","Type":"ContainerStarted","Data":"21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34"} Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.540762 4923 generic.go:334] "Generic (PLEG): container finished" podID="e2bd8524-0fce-4b0d-bffb-0013dac2b65f" containerID="614523fdae72eabb3a0f8ac0f776b5cbfefd10006b8e7fb4436ed44311297c02" exitCode=0 Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.540849 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mx6fg" event={"ID":"e2bd8524-0fce-4b0d-bffb-0013dac2b65f","Type":"ContainerDied","Data":"614523fdae72eabb3a0f8ac0f776b5cbfefd10006b8e7fb4436ed44311297c02"} Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.552337 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hxtts" podStartSLOduration=49.552312535 podStartE2EDuration="49.552312535s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:14.551411973 +0000 UTC m=+102.233313707" watchObservedRunningTime="2026-03-11 13:24:14.552312535 +0000 UTC m=+102.234214279" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.579492 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-rlsnm" podStartSLOduration=49.579471 podStartE2EDuration="49.579471s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:14.574289201 +0000 UTC m=+102.256190925" watchObservedRunningTime="2026-03-11 13:24:14.579471 +0000 UTC m=+102.261372754" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.598932 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fjqzd" podStartSLOduration=49.598914312 podStartE2EDuration="49.598914312s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:14.598855041 +0000 UTC m=+102.280756755" watchObservedRunningTime="2026-03-11 13:24:14.598914312 +0000 UTC m=+102.280816036" Mar 11 13:24:14 crc kubenswrapper[4923]: I0311 13:24:14.819781 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ae7e45f7-8c08-435b-990a-5149da33010a-metrics-certs\") pod \"network-metrics-daemon-4nx54\" (UID: \"ae7e45f7-8c08-435b-990a-5149da33010a\") " pod="openshift-multus/network-metrics-daemon-4nx54" Mar 11 13:24:14 crc kubenswrapper[4923]: E0311 13:24:14.819888 4923 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 11 13:24:14 crc kubenswrapper[4923]: E0311 13:24:14.819947 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ae7e45f7-8c08-435b-990a-5149da33010a-metrics-certs podName:ae7e45f7-8c08-435b-990a-5149da33010a nodeName:}" failed. No retries permitted until 2026-03-11 13:24:16.819933679 +0000 UTC m=+104.501835393 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ae7e45f7-8c08-435b-990a-5149da33010a-metrics-certs") pod "network-metrics-daemon-4nx54" (UID: "ae7e45f7-8c08-435b-990a-5149da33010a") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 11 13:24:15 crc kubenswrapper[4923]: I0311 13:24:15.021792 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4nx54" Mar 11 13:24:15 crc kubenswrapper[4923]: E0311 13:24:15.022002 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4nx54" podUID="ae7e45f7-8c08-435b-990a-5149da33010a" Mar 11 13:24:15 crc kubenswrapper[4923]: I0311 13:24:15.548824 4923 generic.go:334] "Generic (PLEG): container finished" podID="e2bd8524-0fce-4b0d-bffb-0013dac2b65f" containerID="e21c4e2119a9605c00ed2b49810a181dcc85269d7135daf4508fd9a939f7a343" exitCode=0 Mar 11 13:24:15 crc kubenswrapper[4923]: I0311 13:24:15.548927 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mx6fg" event={"ID":"e2bd8524-0fce-4b0d-bffb-0013dac2b65f","Type":"ContainerDied","Data":"e21c4e2119a9605c00ed2b49810a181dcc85269d7135daf4508fd9a939f7a343"} Mar 11 13:24:16 crc kubenswrapper[4923]: I0311 13:24:16.021684 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 11 13:24:16 crc kubenswrapper[4923]: I0311 13:24:16.021837 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 11 13:24:16 crc kubenswrapper[4923]: E0311 13:24:16.022212 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 11 13:24:16 crc kubenswrapper[4923]: I0311 13:24:16.022244 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:16 crc kubenswrapper[4923]: E0311 13:24:16.022441 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 11 13:24:16 crc kubenswrapper[4923]: E0311 13:24:16.022593 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 11 13:24:16 crc kubenswrapper[4923]: I0311 13:24:16.037976 4923 scope.go:117] "RemoveContainer" containerID="9a9f78b1ff5dfa00fd3d80bafc5c131e5fd12de6d0556229087316aad81df459" Mar 11 13:24:16 crc kubenswrapper[4923]: E0311 13:24:16.038287 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 11 13:24:16 crc kubenswrapper[4923]: I0311 13:24:16.040386 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 11 13:24:16 crc kubenswrapper[4923]: I0311 13:24:16.557894 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" event={"ID":"df02f0a6-797a-4c04-a529-f0881241fd94","Type":"ContainerStarted","Data":"436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee"} Mar 11 13:24:16 crc kubenswrapper[4923]: I0311 13:24:16.561445 4923 generic.go:334] "Generic (PLEG): container finished" podID="e2bd8524-0fce-4b0d-bffb-0013dac2b65f" containerID="200251d9c145833c86a475a6841c79db57ae9d7b6a8e37db159cadc022e30a66" exitCode=0 Mar 11 13:24:16 crc kubenswrapper[4923]: I0311 13:24:16.561523 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mx6fg" event={"ID":"e2bd8524-0fce-4b0d-bffb-0013dac2b65f","Type":"ContainerDied","Data":"200251d9c145833c86a475a6841c79db57ae9d7b6a8e37db159cadc022e30a66"} Mar 11 13:24:16 crc kubenswrapper[4923]: I0311 13:24:16.562575 4923 scope.go:117] "RemoveContainer" containerID="9a9f78b1ff5dfa00fd3d80bafc5c131e5fd12de6d0556229087316aad81df459" Mar 11 13:24:16 crc kubenswrapper[4923]: E0311 13:24:16.562776 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 11 13:24:16 crc kubenswrapper[4923]: I0311 13:24:16.843974 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ae7e45f7-8c08-435b-990a-5149da33010a-metrics-certs\") pod \"network-metrics-daemon-4nx54\" (UID: \"ae7e45f7-8c08-435b-990a-5149da33010a\") " pod="openshift-multus/network-metrics-daemon-4nx54" Mar 11 13:24:16 crc kubenswrapper[4923]: E0311 13:24:16.844122 4923 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 11 13:24:16 crc kubenswrapper[4923]: E0311 13:24:16.844218 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ae7e45f7-8c08-435b-990a-5149da33010a-metrics-certs podName:ae7e45f7-8c08-435b-990a-5149da33010a nodeName:}" failed. No retries permitted until 2026-03-11 13:24:20.844194638 +0000 UTC m=+108.526096382 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ae7e45f7-8c08-435b-990a-5149da33010a-metrics-certs") pod "network-metrics-daemon-4nx54" (UID: "ae7e45f7-8c08-435b-990a-5149da33010a") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 11 13:24:17 crc kubenswrapper[4923]: I0311 13:24:17.021406 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4nx54" Mar 11 13:24:17 crc kubenswrapper[4923]: E0311 13:24:17.021562 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4nx54" podUID="ae7e45f7-8c08-435b-990a-5149da33010a" Mar 11 13:24:17 crc kubenswrapper[4923]: I0311 13:24:17.570783 4923 generic.go:334] "Generic (PLEG): container finished" podID="e2bd8524-0fce-4b0d-bffb-0013dac2b65f" containerID="67e6bce66a44eed2b91c557d8c4622151688256214a7ed061d3e1eec0b25c7c7" exitCode=0 Mar 11 13:24:17 crc kubenswrapper[4923]: I0311 13:24:17.570860 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mx6fg" event={"ID":"e2bd8524-0fce-4b0d-bffb-0013dac2b65f","Type":"ContainerDied","Data":"67e6bce66a44eed2b91c557d8c4622151688256214a7ed061d3e1eec0b25c7c7"} Mar 11 13:24:17 crc kubenswrapper[4923]: I0311 13:24:17.757062 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:17 crc kubenswrapper[4923]: E0311 13:24:17.757268 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:25.757236503 +0000 UTC m=+113.439138217 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:17 crc kubenswrapper[4923]: I0311 13:24:17.757476 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:17 crc kubenswrapper[4923]: I0311 13:24:17.757522 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 11 13:24:17 crc kubenswrapper[4923]: I0311 13:24:17.757553 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:17 crc kubenswrapper[4923]: E0311 13:24:17.757606 4923 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 11 13:24:17 crc kubenswrapper[4923]: E0311 13:24:17.757672 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-11 13:24:25.757657275 +0000 UTC m=+113.439559079 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 11 13:24:17 crc kubenswrapper[4923]: E0311 13:24:17.757683 4923 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 11 13:24:17 crc kubenswrapper[4923]: E0311 13:24:17.757740 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-11 13:24:25.757722986 +0000 UTC m=+113.439624710 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 11 13:24:17 crc kubenswrapper[4923]: E0311 13:24:17.757780 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 11 13:24:17 crc kubenswrapper[4923]: E0311 13:24:17.757813 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 11 13:24:17 crc kubenswrapper[4923]: E0311 13:24:17.757826 4923 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 11 13:24:17 crc kubenswrapper[4923]: E0311 13:24:17.757896 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-11 13:24:25.75787686 +0000 UTC m=+113.439778584 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 11 13:24:17 crc kubenswrapper[4923]: I0311 13:24:17.858569 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 11 13:24:17 crc kubenswrapper[4923]: E0311 13:24:17.858742 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 11 13:24:17 crc kubenswrapper[4923]: E0311 13:24:17.858776 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 11 13:24:17 crc kubenswrapper[4923]: E0311 13:24:17.858791 4923 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 11 13:24:17 crc kubenswrapper[4923]: E0311 13:24:17.858860 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-11 13:24:25.858840726 +0000 UTC m=+113.540742450 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 11 13:24:18 crc kubenswrapper[4923]: I0311 13:24:18.020990 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 11 13:24:18 crc kubenswrapper[4923]: I0311 13:24:18.021062 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 11 13:24:18 crc kubenswrapper[4923]: I0311 13:24:18.020990 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:18 crc kubenswrapper[4923]: E0311 13:24:18.021170 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 11 13:24:18 crc kubenswrapper[4923]: E0311 13:24:18.021277 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 11 13:24:18 crc kubenswrapper[4923]: E0311 13:24:18.021475 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 11 13:24:18 crc kubenswrapper[4923]: I0311 13:24:18.579133 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" event={"ID":"df02f0a6-797a-4c04-a529-f0881241fd94","Type":"ContainerStarted","Data":"6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f"} Mar 11 13:24:18 crc kubenswrapper[4923]: I0311 13:24:18.579489 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:18 crc kubenswrapper[4923]: I0311 13:24:18.579507 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:18 crc kubenswrapper[4923]: I0311 13:24:18.579683 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:18 crc kubenswrapper[4923]: I0311 13:24:18.588219 4923 generic.go:334] "Generic (PLEG): container finished" podID="e2bd8524-0fce-4b0d-bffb-0013dac2b65f" containerID="fdc423ff9486627ec3a623ce4c215027d1e788f7bf42c40a3d09a1a94e1c1ab3" exitCode=0 Mar 11 13:24:18 crc kubenswrapper[4923]: I0311 13:24:18.588289 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mx6fg" event={"ID":"e2bd8524-0fce-4b0d-bffb-0013dac2b65f","Type":"ContainerDied","Data":"fdc423ff9486627ec3a623ce4c215027d1e788f7bf42c40a3d09a1a94e1c1ab3"} Mar 11 13:24:18 crc kubenswrapper[4923]: I0311 13:24:18.622044 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" podStartSLOduration=53.622024881 podStartE2EDuration="53.622024881s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:18.62199131 +0000 UTC m=+106.303893034" watchObservedRunningTime="2026-03-11 13:24:18.622024881 +0000 UTC m=+106.303926595" Mar 11 13:24:18 crc kubenswrapper[4923]: I0311 13:24:18.633673 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:18 crc kubenswrapper[4923]: I0311 13:24:18.637609 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:19 crc kubenswrapper[4923]: I0311 13:24:19.021536 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4nx54" Mar 11 13:24:19 crc kubenswrapper[4923]: E0311 13:24:19.021726 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4nx54" podUID="ae7e45f7-8c08-435b-990a-5149da33010a" Mar 11 13:24:19 crc kubenswrapper[4923]: I0311 13:24:19.594286 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mx6fg" event={"ID":"e2bd8524-0fce-4b0d-bffb-0013dac2b65f","Type":"ContainerStarted","Data":"868a1a574b6db7948b2550f1c3e9a581bfb7622a7dc317666e1ed5f187beb478"} Mar 11 13:24:20 crc kubenswrapper[4923]: I0311 13:24:20.020902 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:20 crc kubenswrapper[4923]: I0311 13:24:20.020946 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 11 13:24:20 crc kubenswrapper[4923]: I0311 13:24:20.021045 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 11 13:24:20 crc kubenswrapper[4923]: E0311 13:24:20.021230 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 11 13:24:20 crc kubenswrapper[4923]: E0311 13:24:20.021466 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 11 13:24:20 crc kubenswrapper[4923]: E0311 13:24:20.021613 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 11 13:24:20 crc kubenswrapper[4923]: I0311 13:24:20.675626 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-mx6fg" podStartSLOduration=55.675600828 podStartE2EDuration="55.675600828s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:19.612084408 +0000 UTC m=+107.293986152" watchObservedRunningTime="2026-03-11 13:24:20.675600828 +0000 UTC m=+108.357502582" Mar 11 13:24:20 crc kubenswrapper[4923]: I0311 13:24:20.676608 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-4nx54"] Mar 11 13:24:20 crc kubenswrapper[4923]: I0311 13:24:20.676745 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4nx54" Mar 11 13:24:20 crc kubenswrapper[4923]: E0311 13:24:20.676916 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4nx54" podUID="ae7e45f7-8c08-435b-990a-5149da33010a" Mar 11 13:24:20 crc kubenswrapper[4923]: I0311 13:24:20.894087 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ae7e45f7-8c08-435b-990a-5149da33010a-metrics-certs\") pod \"network-metrics-daemon-4nx54\" (UID: \"ae7e45f7-8c08-435b-990a-5149da33010a\") " pod="openshift-multus/network-metrics-daemon-4nx54" Mar 11 13:24:20 crc kubenswrapper[4923]: E0311 13:24:20.894312 4923 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 11 13:24:20 crc kubenswrapper[4923]: E0311 13:24:20.894502 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ae7e45f7-8c08-435b-990a-5149da33010a-metrics-certs podName:ae7e45f7-8c08-435b-990a-5149da33010a nodeName:}" failed. No retries permitted until 2026-03-11 13:24:28.894467681 +0000 UTC m=+116.576369435 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ae7e45f7-8c08-435b-990a-5149da33010a-metrics-certs") pod "network-metrics-daemon-4nx54" (UID: "ae7e45f7-8c08-435b-990a-5149da33010a") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 11 13:24:22 crc kubenswrapper[4923]: I0311 13:24:22.021563 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 11 13:24:22 crc kubenswrapper[4923]: I0311 13:24:22.021602 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:22 crc kubenswrapper[4923]: E0311 13:24:22.022230 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 11 13:24:22 crc kubenswrapper[4923]: E0311 13:24:22.021993 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 11 13:24:22 crc kubenswrapper[4923]: I0311 13:24:22.021647 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 11 13:24:22 crc kubenswrapper[4923]: E0311 13:24:22.022412 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.021707 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4nx54" Mar 11 13:24:23 crc kubenswrapper[4923]: E0311 13:24:23.023384 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4nx54" podUID="ae7e45f7-8c08-435b-990a-5149da33010a" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.199176 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.199408 4923 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.248057 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-b5jmc"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.248503 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-b5jmc" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.251043 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-4nhkq"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.251366 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-4nhkq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.252848 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.253186 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.254180 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.254312 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r2t4r"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.254760 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r2t4r" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.255853 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-g25ck"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.256146 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.258887 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.259085 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.259436 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.259661 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.259805 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.259998 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.263650 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.263909 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.264093 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.264231 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.266828 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.267101 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.267268 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-c7dks"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.267728 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ndw7p"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.268082 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-ndw7p" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.268458 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-c7dks" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.270551 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8bhpz"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.271253 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8bhpz" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.271793 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.272512 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.275492 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.275602 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-k6dxb"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.276466 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-jfhp5"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.277083 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hgkc2"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.277154 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jfhp5" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.277097 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.278676 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.283216 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.283675 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.283838 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.285145 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-sdtpq"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.295188 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-l6bf9"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.284177 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.284679 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.284753 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.284879 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.295729 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-mgrcg"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.295912 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-sdtpq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.296084 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7rzzf"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.296480 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.312279 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.312722 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.312929 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.313431 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.315408 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.317031 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.319970 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l6bf9" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.346411 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16d21cac-dc41-43fb-a391-8864085617ed-config\") pod \"controller-manager-879f6c89f-g25ck\" (UID: \"16d21cac-dc41-43fb-a391-8864085617ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.346446 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/16d21cac-dc41-43fb-a391-8864085617ed-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-g25ck\" (UID: \"16d21cac-dc41-43fb-a391-8864085617ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.346472 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb5fn\" (UniqueName: \"kubernetes.io/projected/9a054bc8-9c59-47f8-8b37-7eb59a5331a0-kube-api-access-zb5fn\") pod \"machine-api-operator-5694c8668f-c7dks\" (UID: \"9a054bc8-9c59-47f8-8b37-7eb59a5331a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c7dks" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.346496 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f4kq\" (UniqueName: \"kubernetes.io/projected/c7215b86-4f1e-44fb-b0a9-c7d261ce4a54-kube-api-access-2f4kq\") pod \"downloads-7954f5f757-b5jmc\" (UID: \"c7215b86-4f1e-44fb-b0a9-c7d261ce4a54\") " pod="openshift-console/downloads-7954f5f757-b5jmc" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.346519 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9a054bc8-9c59-47f8-8b37-7eb59a5331a0-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-c7dks\" (UID: \"9a054bc8-9c59-47f8-8b37-7eb59a5331a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c7dks" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.346539 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r582n\" (UniqueName: \"kubernetes.io/projected/16d21cac-dc41-43fb-a391-8864085617ed-kube-api-access-r582n\") pod \"controller-manager-879f6c89f-g25ck\" (UID: \"16d21cac-dc41-43fb-a391-8864085617ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.346562 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2w9b\" (UniqueName: \"kubernetes.io/projected/e9b57443-86b7-45e7-a4f7-396719edfc4a-kube-api-access-w2w9b\") pod \"authentication-operator-69f744f599-4nhkq\" (UID: \"e9b57443-86b7-45e7-a4f7-396719edfc4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4nhkq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.346596 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qgkl\" (UniqueName: \"kubernetes.io/projected/34272d37-e3bb-4315-b555-d1b4439657da-kube-api-access-6qgkl\") pod \"dns-operator-744455d44c-ndw7p\" (UID: \"34272d37-e3bb-4315-b555-d1b4439657da\") " pod="openshift-dns-operator/dns-operator-744455d44c-ndw7p" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.346633 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9b57443-86b7-45e7-a4f7-396719edfc4a-serving-cert\") pod \"authentication-operator-69f744f599-4nhkq\" (UID: \"e9b57443-86b7-45e7-a4f7-396719edfc4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4nhkq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.346651 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/34272d37-e3bb-4315-b555-d1b4439657da-metrics-tls\") pod \"dns-operator-744455d44c-ndw7p\" (UID: \"34272d37-e3bb-4315-b555-d1b4439657da\") " pod="openshift-dns-operator/dns-operator-744455d44c-ndw7p" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.346672 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9b57443-86b7-45e7-a4f7-396719edfc4a-config\") pod \"authentication-operator-69f744f599-4nhkq\" (UID: \"e9b57443-86b7-45e7-a4f7-396719edfc4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4nhkq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.346692 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9b57443-86b7-45e7-a4f7-396719edfc4a-service-ca-bundle\") pod \"authentication-operator-69f744f599-4nhkq\" (UID: \"e9b57443-86b7-45e7-a4f7-396719edfc4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4nhkq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.346714 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/48be4455-d57d-412d-9f0d-191b038bcbcb-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-r2t4r\" (UID: \"48be4455-d57d-412d-9f0d-191b038bcbcb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r2t4r" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.346768 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9b57443-86b7-45e7-a4f7-396719edfc4a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-4nhkq\" (UID: \"e9b57443-86b7-45e7-a4f7-396719edfc4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4nhkq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.346787 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16d21cac-dc41-43fb-a391-8864085617ed-serving-cert\") pod \"controller-manager-879f6c89f-g25ck\" (UID: \"16d21cac-dc41-43fb-a391-8864085617ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.346802 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9a054bc8-9c59-47f8-8b37-7eb59a5331a0-images\") pod \"machine-api-operator-5694c8668f-c7dks\" (UID: \"9a054bc8-9c59-47f8-8b37-7eb59a5331a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c7dks" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.346818 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a054bc8-9c59-47f8-8b37-7eb59a5331a0-config\") pod \"machine-api-operator-5694c8668f-c7dks\" (UID: \"9a054bc8-9c59-47f8-8b37-7eb59a5331a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c7dks" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.346846 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/16d21cac-dc41-43fb-a391-8864085617ed-client-ca\") pod \"controller-manager-879f6c89f-g25ck\" (UID: \"16d21cac-dc41-43fb-a391-8864085617ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.346871 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kptnx\" (UniqueName: \"kubernetes.io/projected/48be4455-d57d-412d-9f0d-191b038bcbcb-kube-api-access-kptnx\") pod \"cluster-samples-operator-665b6dd947-r2t4r\" (UID: \"48be4455-d57d-412d-9f0d-191b038bcbcb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r2t4r" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.351781 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.352238 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.352576 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.352831 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.353011 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.353167 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.353415 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.353731 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.353921 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.354069 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.354276 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.354879 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.355008 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.355096 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.352862 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.352896 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.355682 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mm4gf"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.356008 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pk9tb"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.356355 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pk9tb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.356477 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mm4gf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.357421 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.364705 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.364919 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.370365 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.370545 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.370671 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.370773 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.371000 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.371087 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.371156 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.371442 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.371575 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.371864 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.371997 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.372058 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.372074 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.372182 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.372255 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.372324 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.372429 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.372515 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.372553 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.372523 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.372619 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.371608 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.371635 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.372769 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.372880 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.372957 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.373019 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.371578 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.373609 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.373721 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.374187 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-c7dks"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.374265 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.380246 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.380454 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.393281 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.403216 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.403720 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.404030 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.404255 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.404413 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.404480 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.404539 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.404678 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.404677 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.404750 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.404770 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.404812 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.404686 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.404563 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.404886 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.428312 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g96sx"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.429189 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g96sx" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.434131 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.455660 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.456290 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fwfgr"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.456924 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.458113 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.458152 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.458178 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4890044b-8d18-4123-b029-7e2809294be1-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-pk9tb\" (UID: \"4890044b-8d18-4123-b029-7e2809294be1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pk9tb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.458266 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vd2z\" (UniqueName: \"kubernetes.io/projected/7c496315-8928-482f-be56-265af63f3888-kube-api-access-7vd2z\") pod \"etcd-operator-b45778765-7rzzf\" (UID: \"7c496315-8928-482f-be56-265af63f3888\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.459669 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.459895 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.459979 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8bhpz"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.460379 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a054bc8-9c59-47f8-8b37-7eb59a5331a0-config\") pod \"machine-api-operator-5694c8668f-c7dks\" (UID: \"9a054bc8-9c59-47f8-8b37-7eb59a5331a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c7dks" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.460875 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.461002 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-h7jfj"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.461452 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a054bc8-9c59-47f8-8b37-7eb59a5331a0-config\") pod \"machine-api-operator-5694c8668f-c7dks\" (UID: \"9a054bc8-9c59-47f8-8b37-7eb59a5331a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c7dks" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.462476 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.462616 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a67361c9-12cb-4a1a-893c-a735a58da9cd-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.462828 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.462892 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1515573d-f360-41f2-854a-4d0fb437af7d-etcd-client\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.467486 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb7fp\" (UniqueName: \"kubernetes.io/projected/15bb3d2b-b085-4cf9-8ea1-1a67a5bc74e1-kube-api-access-lb7fp\") pod \"openshift-apiserver-operator-796bbdcf4f-mm4gf\" (UID: \"15bb3d2b-b085-4cf9-8ea1-1a67a5bc74e1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mm4gf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.467591 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.467620 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.467648 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6e6cf79-4d28-4262-801f-f85de17c5def-config\") pod \"console-operator-58897d9998-sdtpq\" (UID: \"a6e6cf79-4d28-4262-801f-f85de17c5def\") " pod="openshift-console-operator/console-operator-58897d9998-sdtpq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.467671 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1515573d-f360-41f2-854a-4d0fb437af7d-serving-cert\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.467704 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.467771 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h6lr\" (UniqueName: \"kubernetes.io/projected/1515573d-f360-41f2-854a-4d0fb437af7d-kube-api-access-8h6lr\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.467822 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.467882 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/13238a87-3b9a-471a-8337-7e7f15018221-oauth-serving-cert\") pod \"console-f9d7485db-mgrcg\" (UID: \"13238a87-3b9a-471a-8337-7e7f15018221\") " pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.468079 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/16d21cac-dc41-43fb-a391-8864085617ed-client-ca\") pod \"controller-manager-879f6c89f-g25ck\" (UID: \"16d21cac-dc41-43fb-a391-8864085617ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.469231 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfda8f36-49b9-4081-bbea-f311dd0a87a1-config\") pod \"route-controller-manager-6576b87f9c-qk4ql\" (UID: \"cfda8f36-49b9-4081-bbea-f311dd0a87a1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.469279 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1515573d-f360-41f2-854a-4d0fb437af7d-etcd-serving-ca\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.469309 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ec8d4f8-79c7-4153-94f1-0efe16a3da0f-config\") pod \"machine-approver-56656f9798-jfhp5\" (UID: \"5ec8d4f8-79c7-4153-94f1-0efe16a3da0f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jfhp5" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.469334 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/13238a87-3b9a-471a-8337-7e7f15018221-console-oauth-config\") pod \"console-f9d7485db-mgrcg\" (UID: \"13238a87-3b9a-471a-8337-7e7f15018221\") " pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.469455 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kptnx\" (UniqueName: \"kubernetes.io/projected/48be4455-d57d-412d-9f0d-191b038bcbcb-kube-api-access-kptnx\") pod \"cluster-samples-operator-665b6dd947-r2t4r\" (UID: \"48be4455-d57d-412d-9f0d-191b038bcbcb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r2t4r" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.469479 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/919663b1-49e2-4405-b62a-5cd8b7092953-available-featuregates\") pod \"openshift-config-operator-7777fb866f-l6bf9\" (UID: \"919663b1-49e2-4405-b62a-5cd8b7092953\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l6bf9" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.469508 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcgv5\" (UniqueName: \"kubernetes.io/projected/cfda8f36-49b9-4081-bbea-f311dd0a87a1-kube-api-access-mcgv5\") pod \"route-controller-manager-6576b87f9c-qk4ql\" (UID: \"cfda8f36-49b9-4081-bbea-f311dd0a87a1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.469534 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cfda8f36-49b9-4081-bbea-f311dd0a87a1-client-ca\") pod \"route-controller-manager-6576b87f9c-qk4ql\" (UID: \"cfda8f36-49b9-4081-bbea-f311dd0a87a1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.469558 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cfda8f36-49b9-4081-bbea-f311dd0a87a1-serving-cert\") pod \"route-controller-manager-6576b87f9c-qk4ql\" (UID: \"cfda8f36-49b9-4081-bbea-f311dd0a87a1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.469588 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16d21cac-dc41-43fb-a391-8864085617ed-config\") pod \"controller-manager-879f6c89f-g25ck\" (UID: \"16d21cac-dc41-43fb-a391-8864085617ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.469612 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/16d21cac-dc41-43fb-a391-8864085617ed-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-g25ck\" (UID: \"16d21cac-dc41-43fb-a391-8864085617ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.469633 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-7lx27"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.469907 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h7jfj" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470079 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2c4h7"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.469641 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb5fn\" (UniqueName: \"kubernetes.io/projected/9a054bc8-9c59-47f8-8b37-7eb59a5331a0-kube-api-access-zb5fn\") pod \"machine-api-operator-5694c8668f-c7dks\" (UID: \"9a054bc8-9c59-47f8-8b37-7eb59a5331a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c7dks" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470289 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5ec8d4f8-79c7-4153-94f1-0efe16a3da0f-auth-proxy-config\") pod \"machine-approver-56656f9798-jfhp5\" (UID: \"5ec8d4f8-79c7-4153-94f1-0efe16a3da0f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jfhp5" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470318 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5ec8d4f8-79c7-4153-94f1-0efe16a3da0f-machine-approver-tls\") pod \"machine-approver-56656f9798-jfhp5\" (UID: \"5ec8d4f8-79c7-4153-94f1-0efe16a3da0f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jfhp5" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470410 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470443 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4890044b-8d18-4123-b029-7e2809294be1-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-pk9tb\" (UID: \"4890044b-8d18-4123-b029-7e2809294be1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pk9tb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470482 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f4kq\" (UniqueName: \"kubernetes.io/projected/c7215b86-4f1e-44fb-b0a9-c7d261ce4a54-kube-api-access-2f4kq\") pod \"downloads-7954f5f757-b5jmc\" (UID: \"c7215b86-4f1e-44fb-b0a9-c7d261ce4a54\") " pod="openshift-console/downloads-7954f5f757-b5jmc" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470509 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a6e6cf79-4d28-4262-801f-f85de17c5def-trusted-ca\") pod \"console-operator-58897d9998-sdtpq\" (UID: \"a6e6cf79-4d28-4262-801f-f85de17c5def\") " pod="openshift-console-operator/console-operator-58897d9998-sdtpq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470547 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f4826a3f-d90e-4b7c-a10b-864b78115463-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8bhpz\" (UID: \"f4826a3f-d90e-4b7c-a10b-864b78115463\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8bhpz" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470583 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9a054bc8-9c59-47f8-8b37-7eb59a5331a0-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-c7dks\" (UID: \"9a054bc8-9c59-47f8-8b37-7eb59a5331a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c7dks" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470609 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmsbn\" (UniqueName: \"kubernetes.io/projected/a6e6cf79-4d28-4262-801f-f85de17c5def-kube-api-access-pmsbn\") pod \"console-operator-58897d9998-sdtpq\" (UID: \"a6e6cf79-4d28-4262-801f-f85de17c5def\") " pod="openshift-console-operator/console-operator-58897d9998-sdtpq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470646 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r582n\" (UniqueName: \"kubernetes.io/projected/16d21cac-dc41-43fb-a391-8864085617ed-kube-api-access-r582n\") pod \"controller-manager-879f6c89f-g25ck\" (UID: \"16d21cac-dc41-43fb-a391-8864085617ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470673 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470700 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2c4h7" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470711 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a67361c9-12cb-4a1a-893c-a735a58da9cd-serving-cert\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470737 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a67361c9-12cb-4a1a-893c-a735a58da9cd-encryption-config\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470761 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7c496315-8928-482f-be56-265af63f3888-etcd-service-ca\") pod \"etcd-operator-b45778765-7rzzf\" (UID: \"7c496315-8928-482f-be56-265af63f3888\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470783 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1515573d-f360-41f2-854a-4d0fb437af7d-encryption-config\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470808 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/13238a87-3b9a-471a-8337-7e7f15018221-console-config\") pod \"console-f9d7485db-mgrcg\" (UID: \"13238a87-3b9a-471a-8337-7e7f15018221\") " pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470833 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jknjt\" (UniqueName: \"kubernetes.io/projected/919663b1-49e2-4405-b62a-5cd8b7092953-kube-api-access-jknjt\") pod \"openshift-config-operator-7777fb866f-l6bf9\" (UID: \"919663b1-49e2-4405-b62a-5cd8b7092953\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l6bf9" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470866 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2w9b\" (UniqueName: \"kubernetes.io/projected/e9b57443-86b7-45e7-a4f7-396719edfc4a-kube-api-access-w2w9b\") pod \"authentication-operator-69f744f599-4nhkq\" (UID: \"e9b57443-86b7-45e7-a4f7-396719edfc4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4nhkq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470891 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-7lx27" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470893 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qgkl\" (UniqueName: \"kubernetes.io/projected/34272d37-e3bb-4315-b555-d1b4439657da-kube-api-access-6qgkl\") pod \"dns-operator-744455d44c-ndw7p\" (UID: \"34272d37-e3bb-4315-b555-d1b4439657da\") " pod="openshift-dns-operator/dns-operator-744455d44c-ndw7p" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470936 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a67361c9-12cb-4a1a-893c-a735a58da9cd-audit-dir\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.470985 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15bb3d2b-b085-4cf9-8ea1-1a67a5bc74e1-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-mm4gf\" (UID: \"15bb3d2b-b085-4cf9-8ea1-1a67a5bc74e1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mm4gf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.471006 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1515573d-f360-41f2-854a-4d0fb437af7d-config\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.471027 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9b57443-86b7-45e7-a4f7-396719edfc4a-serving-cert\") pod \"authentication-operator-69f744f599-4nhkq\" (UID: \"e9b57443-86b7-45e7-a4f7-396719edfc4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4nhkq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.471045 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/34272d37-e3bb-4315-b555-d1b4439657da-metrics-tls\") pod \"dns-operator-744455d44c-ndw7p\" (UID: \"34272d37-e3bb-4315-b555-d1b4439657da\") " pod="openshift-dns-operator/dns-operator-744455d44c-ndw7p" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.471063 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/13238a87-3b9a-471a-8337-7e7f15018221-console-serving-cert\") pod \"console-f9d7485db-mgrcg\" (UID: \"13238a87-3b9a-471a-8337-7e7f15018221\") " pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.471082 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c496315-8928-482f-be56-265af63f3888-config\") pod \"etcd-operator-b45778765-7rzzf\" (UID: \"7c496315-8928-482f-be56-265af63f3888\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.471104 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1515573d-f360-41f2-854a-4d0fb437af7d-node-pullsecrets\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.471127 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16d21cac-dc41-43fb-a391-8864085617ed-config\") pod \"controller-manager-879f6c89f-g25ck\" (UID: \"16d21cac-dc41-43fb-a391-8864085617ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.471144 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9b57443-86b7-45e7-a4f7-396719edfc4a-config\") pod \"authentication-operator-69f744f599-4nhkq\" (UID: \"e9b57443-86b7-45e7-a4f7-396719edfc4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4nhkq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.471175 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9b57443-86b7-45e7-a4f7-396719edfc4a-service-ca-bundle\") pod \"authentication-operator-69f744f599-4nhkq\" (UID: \"e9b57443-86b7-45e7-a4f7-396719edfc4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4nhkq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.471476 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/48be4455-d57d-412d-9f0d-191b038bcbcb-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-r2t4r\" (UID: \"48be4455-d57d-412d-9f0d-191b038bcbcb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r2t4r" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.471506 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a67361c9-12cb-4a1a-893c-a735a58da9cd-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.471524 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1515573d-f360-41f2-854a-4d0fb437af7d-audit\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.471544 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1515573d-f360-41f2-854a-4d0fb437af7d-audit-dir\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.471578 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm2p2\" (UniqueName: \"kubernetes.io/projected/f4826a3f-d90e-4b7c-a10b-864b78115463-kube-api-access-gm2p2\") pod \"cluster-image-registry-operator-dc59b4c8b-8bhpz\" (UID: \"f4826a3f-d90e-4b7c-a10b-864b78115463\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8bhpz" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.471597 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a67361c9-12cb-4a1a-893c-a735a58da9cd-etcd-client\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.471612 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-audit-policies\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.471611 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/16d21cac-dc41-43fb-a391-8864085617ed-client-ca\") pod \"controller-manager-879f6c89f-g25ck\" (UID: \"16d21cac-dc41-43fb-a391-8864085617ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.471633 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvrdv\" (UniqueName: \"kubernetes.io/projected/a67361c9-12cb-4a1a-893c-a735a58da9cd-kube-api-access-zvrdv\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.471648 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c496315-8928-482f-be56-265af63f3888-serving-cert\") pod \"etcd-operator-b45778765-7rzzf\" (UID: \"7c496315-8928-482f-be56-265af63f3888\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.471664 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1515573d-f360-41f2-854a-4d0fb437af7d-image-import-ca\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.471707 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbvxd\" (UniqueName: \"kubernetes.io/projected/13238a87-3b9a-471a-8337-7e7f15018221-kube-api-access-xbvxd\") pod \"console-f9d7485db-mgrcg\" (UID: \"13238a87-3b9a-471a-8337-7e7f15018221\") " pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.471724 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/919663b1-49e2-4405-b62a-5cd8b7092953-serving-cert\") pod \"openshift-config-operator-7777fb866f-l6bf9\" (UID: \"919663b1-49e2-4405-b62a-5cd8b7092953\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l6bf9" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.472829 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/16d21cac-dc41-43fb-a391-8864085617ed-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-g25ck\" (UID: \"16d21cac-dc41-43fb-a391-8864085617ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.473598 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.476256 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qnljg"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.476888 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-zd9ns"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.477993 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a67361c9-12cb-4a1a-893c-a735a58da9cd-audit-policies\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.478597 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9b57443-86b7-45e7-a4f7-396719edfc4a-config\") pod \"authentication-operator-69f744f599-4nhkq\" (UID: \"e9b57443-86b7-45e7-a4f7-396719edfc4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4nhkq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.479306 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.479383 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/13238a87-3b9a-471a-8337-7e7f15018221-trusted-ca-bundle\") pod \"console-f9d7485db-mgrcg\" (UID: \"13238a87-3b9a-471a-8337-7e7f15018221\") " pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.479496 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/08068104-34d8-440d-8f98-f9a6216af858-audit-dir\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.479551 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm9c2\" (UniqueName: \"kubernetes.io/projected/5ec8d4f8-79c7-4153-94f1-0efe16a3da0f-kube-api-access-wm9c2\") pod \"machine-approver-56656f9798-jfhp5\" (UID: \"5ec8d4f8-79c7-4153-94f1-0efe16a3da0f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jfhp5" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.479592 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.479621 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/13238a87-3b9a-471a-8337-7e7f15018221-service-ca\") pod \"console-f9d7485db-mgrcg\" (UID: \"13238a87-3b9a-471a-8337-7e7f15018221\") " pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.479683 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7c496315-8928-482f-be56-265af63f3888-etcd-client\") pod \"etcd-operator-b45778765-7rzzf\" (UID: \"7c496315-8928-482f-be56-265af63f3888\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.479715 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6e6cf79-4d28-4262-801f-f85de17c5def-serving-cert\") pod \"console-operator-58897d9998-sdtpq\" (UID: \"a6e6cf79-4d28-4262-801f-f85de17c5def\") " pod="openshift-console-operator/console-operator-58897d9998-sdtpq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.479744 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbktl\" (UniqueName: \"kubernetes.io/projected/4890044b-8d18-4123-b029-7e2809294be1-kube-api-access-qbktl\") pod \"openshift-controller-manager-operator-756b6f6bc6-pk9tb\" (UID: \"4890044b-8d18-4123-b029-7e2809294be1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pk9tb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.479801 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f4826a3f-d90e-4b7c-a10b-864b78115463-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8bhpz\" (UID: \"f4826a3f-d90e-4b7c-a10b-864b78115463\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8bhpz" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.479825 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx5b4\" (UniqueName: \"kubernetes.io/projected/08068104-34d8-440d-8f98-f9a6216af858-kube-api-access-nx5b4\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.479899 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9b57443-86b7-45e7-a4f7-396719edfc4a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-4nhkq\" (UID: \"e9b57443-86b7-45e7-a4f7-396719edfc4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4nhkq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.479974 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7c496315-8928-482f-be56-265af63f3888-etcd-ca\") pod \"etcd-operator-b45778765-7rzzf\" (UID: \"7c496315-8928-482f-be56-265af63f3888\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.480015 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1515573d-f360-41f2-854a-4d0fb437af7d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.480059 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.480068 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16d21cac-dc41-43fb-a391-8864085617ed-serving-cert\") pod \"controller-manager-879f6c89f-g25ck\" (UID: \"16d21cac-dc41-43fb-a391-8864085617ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.480377 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-zd9ns" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.480778 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.480867 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qnljg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.481546 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9b57443-86b7-45e7-a4f7-396719edfc4a-service-ca-bundle\") pod \"authentication-operator-69f744f599-4nhkq\" (UID: \"e9b57443-86b7-45e7-a4f7-396719edfc4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4nhkq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.482520 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9a054bc8-9c59-47f8-8b37-7eb59a5331a0-images\") pod \"machine-api-operator-5694c8668f-c7dks\" (UID: \"9a054bc8-9c59-47f8-8b37-7eb59a5331a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c7dks" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.482557 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f4826a3f-d90e-4b7c-a10b-864b78115463-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8bhpz\" (UID: \"f4826a3f-d90e-4b7c-a10b-864b78115463\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8bhpz" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.482581 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15bb3d2b-b085-4cf9-8ea1-1a67a5bc74e1-config\") pod \"openshift-apiserver-operator-796bbdcf4f-mm4gf\" (UID: \"15bb3d2b-b085-4cf9-8ea1-1a67a5bc74e1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mm4gf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.483261 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.483393 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16d21cac-dc41-43fb-a391-8864085617ed-serving-cert\") pod \"controller-manager-879f6c89f-g25ck\" (UID: \"16d21cac-dc41-43fb-a391-8864085617ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.483416 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9grj8"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.486370 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-9c5lc"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.485751 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e9b57443-86b7-45e7-a4f7-396719edfc4a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-4nhkq\" (UID: \"e9b57443-86b7-45e7-a4f7-396719edfc4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4nhkq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.486722 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9a054bc8-9c59-47f8-8b37-7eb59a5331a0-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-c7dks\" (UID: \"9a054bc8-9c59-47f8-8b37-7eb59a5331a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c7dks" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.483439 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9a054bc8-9c59-47f8-8b37-7eb59a5331a0-images\") pod \"machine-api-operator-5694c8668f-c7dks\" (UID: \"9a054bc8-9c59-47f8-8b37-7eb59a5331a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c7dks" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.486802 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/34272d37-e3bb-4315-b555-d1b4439657da-metrics-tls\") pod \"dns-operator-744455d44c-ndw7p\" (UID: \"34272d37-e3bb-4315-b555-d1b4439657da\") " pod="openshift-dns-operator/dns-operator-744455d44c-ndw7p" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.487065 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9grj8" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.487992 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6v8k4"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.488189 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-9c5lc" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.490509 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/48be4455-d57d-412d-9f0d-191b038bcbcb-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-r2t4r\" (UID: \"48be4455-d57d-412d-9f0d-191b038bcbcb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r2t4r" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.491668 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6v8k4" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.493568 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.493850 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.493979 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.495109 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqr7s"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.495898 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.495916 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqr7s" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.496623 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.496780 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.499269 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pk9tb"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.500692 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6jmqc"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.501086 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e9b57443-86b7-45e7-a4f7-396719edfc4a-serving-cert\") pod \"authentication-operator-69f744f599-4nhkq\" (UID: \"e9b57443-86b7-45e7-a4f7-396719edfc4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4nhkq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.501608 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-6jmqc" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.502623 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.503232 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.504180 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29553915-5tqf9"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.504769 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29553915-5tqf9" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.505319 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5r4gt"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.505892 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5r4gt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.506455 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9cjnl"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.508837 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-4nhkq"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.509215 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9cjnl" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.511489 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-lb2kf"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.515298 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.515794 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h59nb"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.516054 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lb2kf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.523623 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g96sx"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.523682 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-dxlk5"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.524402 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h59nb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.524900 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mm4gf"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.524928 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vznqg"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.525178 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dxlk5" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.525639 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.527057 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qkhqm"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.528108 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qkhqm" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.529032 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-l6bf9"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.530284 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-g25ck"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.531138 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.531645 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hgkc2"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.533434 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-sdtpq"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.534222 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-mgrcg"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.535508 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-57kdl"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.536437 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.536816 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-k6dxb"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.538903 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-wdt5g"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.541705 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fwfgr"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.541728 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-h7jfj"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.541799 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-wdt5g" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.541941 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-9c5lc"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.542984 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-7lx27"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.544238 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2c4h7"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.545437 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ndw7p"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.546742 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqr7s"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.547780 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qnljg"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.548811 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6jmqc"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.550200 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h59nb"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.551055 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.551205 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9grj8"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.552507 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r2t4r"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.553543 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-b5jmc"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.554583 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.555642 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-lb2kf"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.556475 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29553915-5tqf9"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.557429 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6v8k4"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.558391 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.559494 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qkhqm"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.560360 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-dxlk5"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.561255 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5r4gt"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.562200 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7rzzf"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.563060 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-lc9xf"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.564176 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-m4jm6"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.564385 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.565665 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-wdt5g"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.565759 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-m4jm6" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.566375 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9cjnl"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.567198 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vznqg"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.568088 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-lc9xf"] Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.571787 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.583706 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfda8f36-49b9-4081-bbea-f311dd0a87a1-config\") pod \"route-controller-manager-6576b87f9c-qk4ql\" (UID: \"cfda8f36-49b9-4081-bbea-f311dd0a87a1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.583744 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1515573d-f360-41f2-854a-4d0fb437af7d-etcd-serving-ca\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.583773 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ec8d4f8-79c7-4153-94f1-0efe16a3da0f-config\") pod \"machine-approver-56656f9798-jfhp5\" (UID: \"5ec8d4f8-79c7-4153-94f1-0efe16a3da0f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jfhp5" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.583797 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/13238a87-3b9a-471a-8337-7e7f15018221-console-oauth-config\") pod \"console-f9d7485db-mgrcg\" (UID: \"13238a87-3b9a-471a-8337-7e7f15018221\") " pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.583844 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/919663b1-49e2-4405-b62a-5cd8b7092953-available-featuregates\") pod \"openshift-config-operator-7777fb866f-l6bf9\" (UID: \"919663b1-49e2-4405-b62a-5cd8b7092953\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l6bf9" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.583869 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcgv5\" (UniqueName: \"kubernetes.io/projected/cfda8f36-49b9-4081-bbea-f311dd0a87a1-kube-api-access-mcgv5\") pod \"route-controller-manager-6576b87f9c-qk4ql\" (UID: \"cfda8f36-49b9-4081-bbea-f311dd0a87a1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.583890 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cfda8f36-49b9-4081-bbea-f311dd0a87a1-client-ca\") pod \"route-controller-manager-6576b87f9c-qk4ql\" (UID: \"cfda8f36-49b9-4081-bbea-f311dd0a87a1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.583915 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cfda8f36-49b9-4081-bbea-f311dd0a87a1-serving-cert\") pod \"route-controller-manager-6576b87f9c-qk4ql\" (UID: \"cfda8f36-49b9-4081-bbea-f311dd0a87a1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.583937 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4890044b-8d18-4123-b029-7e2809294be1-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-pk9tb\" (UID: \"4890044b-8d18-4123-b029-7e2809294be1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pk9tb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.583966 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5ec8d4f8-79c7-4153-94f1-0efe16a3da0f-auth-proxy-config\") pod \"machine-approver-56656f9798-jfhp5\" (UID: \"5ec8d4f8-79c7-4153-94f1-0efe16a3da0f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jfhp5" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.583986 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5ec8d4f8-79c7-4153-94f1-0efe16a3da0f-machine-approver-tls\") pod \"machine-approver-56656f9798-jfhp5\" (UID: \"5ec8d4f8-79c7-4153-94f1-0efe16a3da0f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jfhp5" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584011 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584046 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a6e6cf79-4d28-4262-801f-f85de17c5def-trusted-ca\") pod \"console-operator-58897d9998-sdtpq\" (UID: \"a6e6cf79-4d28-4262-801f-f85de17c5def\") " pod="openshift-console-operator/console-operator-58897d9998-sdtpq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584072 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f4826a3f-d90e-4b7c-a10b-864b78115463-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8bhpz\" (UID: \"f4826a3f-d90e-4b7c-a10b-864b78115463\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8bhpz" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584099 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmsbn\" (UniqueName: \"kubernetes.io/projected/a6e6cf79-4d28-4262-801f-f85de17c5def-kube-api-access-pmsbn\") pod \"console-operator-58897d9998-sdtpq\" (UID: \"a6e6cf79-4d28-4262-801f-f85de17c5def\") " pod="openshift-console-operator/console-operator-58897d9998-sdtpq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584134 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584158 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1515573d-f360-41f2-854a-4d0fb437af7d-encryption-config\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584183 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a67361c9-12cb-4a1a-893c-a735a58da9cd-serving-cert\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584205 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a67361c9-12cb-4a1a-893c-a735a58da9cd-encryption-config\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584227 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7c496315-8928-482f-be56-265af63f3888-etcd-service-ca\") pod \"etcd-operator-b45778765-7rzzf\" (UID: \"7c496315-8928-482f-be56-265af63f3888\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584251 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/13238a87-3b9a-471a-8337-7e7f15018221-console-config\") pod \"console-f9d7485db-mgrcg\" (UID: \"13238a87-3b9a-471a-8337-7e7f15018221\") " pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584273 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jknjt\" (UniqueName: \"kubernetes.io/projected/919663b1-49e2-4405-b62a-5cd8b7092953-kube-api-access-jknjt\") pod \"openshift-config-operator-7777fb866f-l6bf9\" (UID: \"919663b1-49e2-4405-b62a-5cd8b7092953\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l6bf9" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584283 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/919663b1-49e2-4405-b62a-5cd8b7092953-available-featuregates\") pod \"openshift-config-operator-7777fb866f-l6bf9\" (UID: \"919663b1-49e2-4405-b62a-5cd8b7092953\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l6bf9" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584312 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a67361c9-12cb-4a1a-893c-a735a58da9cd-audit-dir\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584325 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ec8d4f8-79c7-4153-94f1-0efe16a3da0f-config\") pod \"machine-approver-56656f9798-jfhp5\" (UID: \"5ec8d4f8-79c7-4153-94f1-0efe16a3da0f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jfhp5" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584374 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1515573d-f360-41f2-854a-4d0fb437af7d-config\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584416 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15bb3d2b-b085-4cf9-8ea1-1a67a5bc74e1-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-mm4gf\" (UID: \"15bb3d2b-b085-4cf9-8ea1-1a67a5bc74e1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mm4gf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584445 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/13238a87-3b9a-471a-8337-7e7f15018221-console-serving-cert\") pod \"console-f9d7485db-mgrcg\" (UID: \"13238a87-3b9a-471a-8337-7e7f15018221\") " pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584463 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c496315-8928-482f-be56-265af63f3888-config\") pod \"etcd-operator-b45778765-7rzzf\" (UID: \"7c496315-8928-482f-be56-265af63f3888\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584510 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1515573d-f360-41f2-854a-4d0fb437af7d-node-pullsecrets\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584536 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a67361c9-12cb-4a1a-893c-a735a58da9cd-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584551 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1515573d-f360-41f2-854a-4d0fb437af7d-audit\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584569 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1515573d-f360-41f2-854a-4d0fb437af7d-audit-dir\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584597 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm2p2\" (UniqueName: \"kubernetes.io/projected/f4826a3f-d90e-4b7c-a10b-864b78115463-kube-api-access-gm2p2\") pod \"cluster-image-registry-operator-dc59b4c8b-8bhpz\" (UID: \"f4826a3f-d90e-4b7c-a10b-864b78115463\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8bhpz" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584615 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a67361c9-12cb-4a1a-893c-a735a58da9cd-etcd-client\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584633 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-audit-policies\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584653 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvrdv\" (UniqueName: \"kubernetes.io/projected/a67361c9-12cb-4a1a-893c-a735a58da9cd-kube-api-access-zvrdv\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584669 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c496315-8928-482f-be56-265af63f3888-serving-cert\") pod \"etcd-operator-b45778765-7rzzf\" (UID: \"7c496315-8928-482f-be56-265af63f3888\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584684 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1515573d-f360-41f2-854a-4d0fb437af7d-image-import-ca\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584701 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbvxd\" (UniqueName: \"kubernetes.io/projected/13238a87-3b9a-471a-8337-7e7f15018221-kube-api-access-xbvxd\") pod \"console-f9d7485db-mgrcg\" (UID: \"13238a87-3b9a-471a-8337-7e7f15018221\") " pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584721 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/919663b1-49e2-4405-b62a-5cd8b7092953-serving-cert\") pod \"openshift-config-operator-7777fb866f-l6bf9\" (UID: \"919663b1-49e2-4405-b62a-5cd8b7092953\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l6bf9" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584737 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a67361c9-12cb-4a1a-893c-a735a58da9cd-audit-policies\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584756 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584770 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/13238a87-3b9a-471a-8337-7e7f15018221-trusted-ca-bundle\") pod \"console-f9d7485db-mgrcg\" (UID: \"13238a87-3b9a-471a-8337-7e7f15018221\") " pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584776 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1515573d-f360-41f2-854a-4d0fb437af7d-etcd-serving-ca\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584787 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/08068104-34d8-440d-8f98-f9a6216af858-audit-dir\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584807 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm9c2\" (UniqueName: \"kubernetes.io/projected/5ec8d4f8-79c7-4153-94f1-0efe16a3da0f-kube-api-access-wm9c2\") pod \"machine-approver-56656f9798-jfhp5\" (UID: \"5ec8d4f8-79c7-4153-94f1-0efe16a3da0f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jfhp5" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584827 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584844 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/13238a87-3b9a-471a-8337-7e7f15018221-service-ca\") pod \"console-f9d7485db-mgrcg\" (UID: \"13238a87-3b9a-471a-8337-7e7f15018221\") " pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584858 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7c496315-8928-482f-be56-265af63f3888-etcd-client\") pod \"etcd-operator-b45778765-7rzzf\" (UID: \"7c496315-8928-482f-be56-265af63f3888\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584875 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6e6cf79-4d28-4262-801f-f85de17c5def-serving-cert\") pod \"console-operator-58897d9998-sdtpq\" (UID: \"a6e6cf79-4d28-4262-801f-f85de17c5def\") " pod="openshift-console-operator/console-operator-58897d9998-sdtpq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584893 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbktl\" (UniqueName: \"kubernetes.io/projected/4890044b-8d18-4123-b029-7e2809294be1-kube-api-access-qbktl\") pod \"openshift-controller-manager-operator-756b6f6bc6-pk9tb\" (UID: \"4890044b-8d18-4123-b029-7e2809294be1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pk9tb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584904 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1515573d-f360-41f2-854a-4d0fb437af7d-config\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584915 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f4826a3f-d90e-4b7c-a10b-864b78115463-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8bhpz\" (UID: \"f4826a3f-d90e-4b7c-a10b-864b78115463\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8bhpz" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584931 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx5b4\" (UniqueName: \"kubernetes.io/projected/08068104-34d8-440d-8f98-f9a6216af858-kube-api-access-nx5b4\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584951 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7c496315-8928-482f-be56-265af63f3888-etcd-ca\") pod \"etcd-operator-b45778765-7rzzf\" (UID: \"7c496315-8928-482f-be56-265af63f3888\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584968 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1515573d-f360-41f2-854a-4d0fb437af7d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584996 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f4826a3f-d90e-4b7c-a10b-864b78115463-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8bhpz\" (UID: \"f4826a3f-d90e-4b7c-a10b-864b78115463\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8bhpz" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.585014 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15bb3d2b-b085-4cf9-8ea1-1a67a5bc74e1-config\") pod \"openshift-apiserver-operator-796bbdcf4f-mm4gf\" (UID: \"15bb3d2b-b085-4cf9-8ea1-1a67a5bc74e1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mm4gf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.585032 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.585052 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.585071 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4890044b-8d18-4123-b029-7e2809294be1-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-pk9tb\" (UID: \"4890044b-8d18-4123-b029-7e2809294be1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pk9tb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.585087 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vd2z\" (UniqueName: \"kubernetes.io/projected/7c496315-8928-482f-be56-265af63f3888-kube-api-access-7vd2z\") pod \"etcd-operator-b45778765-7rzzf\" (UID: \"7c496315-8928-482f-be56-265af63f3888\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.585114 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a67361c9-12cb-4a1a-893c-a735a58da9cd-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.585130 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.585146 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1515573d-f360-41f2-854a-4d0fb437af7d-etcd-client\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.585163 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lb7fp\" (UniqueName: \"kubernetes.io/projected/15bb3d2b-b085-4cf9-8ea1-1a67a5bc74e1-kube-api-access-lb7fp\") pod \"openshift-apiserver-operator-796bbdcf4f-mm4gf\" (UID: \"15bb3d2b-b085-4cf9-8ea1-1a67a5bc74e1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mm4gf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.585180 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.585197 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.585212 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6e6cf79-4d28-4262-801f-f85de17c5def-config\") pod \"console-operator-58897d9998-sdtpq\" (UID: \"a6e6cf79-4d28-4262-801f-f85de17c5def\") " pod="openshift-console-operator/console-operator-58897d9998-sdtpq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.585225 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1515573d-f360-41f2-854a-4d0fb437af7d-serving-cert\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.585242 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.585258 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h6lr\" (UniqueName: \"kubernetes.io/projected/1515573d-f360-41f2-854a-4d0fb437af7d-kube-api-access-8h6lr\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.585266 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a67361c9-12cb-4a1a-893c-a735a58da9cd-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.585284 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.585284 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4890044b-8d18-4123-b029-7e2809294be1-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-pk9tb\" (UID: \"4890044b-8d18-4123-b029-7e2809294be1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pk9tb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.585301 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/13238a87-3b9a-471a-8337-7e7f15018221-oauth-serving-cert\") pod \"console-f9d7485db-mgrcg\" (UID: \"13238a87-3b9a-471a-8337-7e7f15018221\") " pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.585944 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c496315-8928-482f-be56-265af63f3888-config\") pod \"etcd-operator-b45778765-7rzzf\" (UID: \"7c496315-8928-482f-be56-265af63f3888\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.585954 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cfda8f36-49b9-4081-bbea-f311dd0a87a1-client-ca\") pod \"route-controller-manager-6576b87f9c-qk4ql\" (UID: \"cfda8f36-49b9-4081-bbea-f311dd0a87a1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.585994 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/13238a87-3b9a-471a-8337-7e7f15018221-oauth-serving-cert\") pod \"console-f9d7485db-mgrcg\" (UID: \"13238a87-3b9a-471a-8337-7e7f15018221\") " pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.586036 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/1515573d-f360-41f2-854a-4d0fb437af7d-node-pullsecrets\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.584377 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a67361c9-12cb-4a1a-893c-a735a58da9cd-audit-dir\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.596592 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a67361c9-12cb-4a1a-893c-a735a58da9cd-audit-policies\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.596740 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfda8f36-49b9-4081-bbea-f311dd0a87a1-config\") pod \"route-controller-manager-6576b87f9c-qk4ql\" (UID: \"cfda8f36-49b9-4081-bbea-f311dd0a87a1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.596765 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5ec8d4f8-79c7-4153-94f1-0efe16a3da0f-auth-proxy-config\") pod \"machine-approver-56656f9798-jfhp5\" (UID: \"5ec8d4f8-79c7-4153-94f1-0efe16a3da0f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jfhp5" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.597161 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/13238a87-3b9a-471a-8337-7e7f15018221-console-serving-cert\") pod \"console-f9d7485db-mgrcg\" (UID: \"13238a87-3b9a-471a-8337-7e7f15018221\") " pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.597238 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cfda8f36-49b9-4081-bbea-f311dd0a87a1-serving-cert\") pod \"route-controller-manager-6576b87f9c-qk4ql\" (UID: \"cfda8f36-49b9-4081-bbea-f311dd0a87a1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.597264 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/13238a87-3b9a-471a-8337-7e7f15018221-console-oauth-config\") pod \"console-f9d7485db-mgrcg\" (UID: \"13238a87-3b9a-471a-8337-7e7f15018221\") " pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.597486 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.597659 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1515573d-f360-41f2-854a-4d0fb437af7d-audit-dir\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.597658 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.597751 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1515573d-f360-41f2-854a-4d0fb437af7d-audit\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.597784 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-audit-policies\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.598114 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7c496315-8928-482f-be56-265af63f3888-etcd-service-ca\") pod \"etcd-operator-b45778765-7rzzf\" (UID: \"7c496315-8928-482f-be56-265af63f3888\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.599286 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15bb3d2b-b085-4cf9-8ea1-1a67a5bc74e1-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-mm4gf\" (UID: \"15bb3d2b-b085-4cf9-8ea1-1a67a5bc74e1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mm4gf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.599478 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1515573d-f360-41f2-854a-4d0fb437af7d-encryption-config\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.599902 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/13238a87-3b9a-471a-8337-7e7f15018221-console-config\") pod \"console-f9d7485db-mgrcg\" (UID: \"13238a87-3b9a-471a-8337-7e7f15018221\") " pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.600079 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/919663b1-49e2-4405-b62a-5cd8b7092953-serving-cert\") pod \"openshift-config-operator-7777fb866f-l6bf9\" (UID: \"919663b1-49e2-4405-b62a-5cd8b7092953\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l6bf9" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.600246 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.600396 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1515573d-f360-41f2-854a-4d0fb437af7d-serving-cert\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.600599 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/5ec8d4f8-79c7-4153-94f1-0efe16a3da0f-machine-approver-tls\") pod \"machine-approver-56656f9798-jfhp5\" (UID: \"5ec8d4f8-79c7-4153-94f1-0efe16a3da0f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jfhp5" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.601695 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a67361c9-12cb-4a1a-893c-a735a58da9cd-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.602935 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/f4826a3f-d90e-4b7c-a10b-864b78115463-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-8bhpz\" (UID: \"f4826a3f-d90e-4b7c-a10b-864b78115463\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8bhpz" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.603592 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/08068104-34d8-440d-8f98-f9a6216af858-audit-dir\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.603950 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/13238a87-3b9a-471a-8337-7e7f15018221-service-ca\") pod \"console-f9d7485db-mgrcg\" (UID: \"13238a87-3b9a-471a-8337-7e7f15018221\") " pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.604974 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4890044b-8d18-4123-b029-7e2809294be1-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-pk9tb\" (UID: \"4890044b-8d18-4123-b029-7e2809294be1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pk9tb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.605191 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a6e6cf79-4d28-4262-801f-f85de17c5def-trusted-ca\") pod \"console-operator-58897d9998-sdtpq\" (UID: \"a6e6cf79-4d28-4262-801f-f85de17c5def\") " pod="openshift-console-operator/console-operator-58897d9998-sdtpq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.605222 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.605996 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/13238a87-3b9a-471a-8337-7e7f15018221-trusted-ca-bundle\") pod \"console-f9d7485db-mgrcg\" (UID: \"13238a87-3b9a-471a-8337-7e7f15018221\") " pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.606137 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7c496315-8928-482f-be56-265af63f3888-etcd-ca\") pod \"etcd-operator-b45778765-7rzzf\" (UID: \"7c496315-8928-482f-be56-265af63f3888\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.606316 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1515573d-f360-41f2-854a-4d0fb437af7d-trusted-ca-bundle\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.606464 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6e6cf79-4d28-4262-801f-f85de17c5def-config\") pod \"console-operator-58897d9998-sdtpq\" (UID: \"a6e6cf79-4d28-4262-801f-f85de17c5def\") " pod="openshift-console-operator/console-operator-58897d9998-sdtpq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.606889 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.607031 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a67361c9-12cb-4a1a-893c-a735a58da9cd-encryption-config\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.607201 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a67361c9-12cb-4a1a-893c-a735a58da9cd-etcd-client\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.607500 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7c496315-8928-482f-be56-265af63f3888-etcd-client\") pod \"etcd-operator-b45778765-7rzzf\" (UID: \"7c496315-8928-482f-be56-265af63f3888\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.607724 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1515573d-f360-41f2-854a-4d0fb437af7d-image-import-ca\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.607678 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.607836 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.607632 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.607894 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a67361c9-12cb-4a1a-893c-a735a58da9cd-serving-cert\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.608482 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f4826a3f-d90e-4b7c-a10b-864b78115463-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-8bhpz\" (UID: \"f4826a3f-d90e-4b7c-a10b-864b78115463\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8bhpz" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.609192 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7c496315-8928-482f-be56-265af63f3888-serving-cert\") pod \"etcd-operator-b45778765-7rzzf\" (UID: \"7c496315-8928-482f-be56-265af63f3888\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.609252 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.609539 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.609566 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1515573d-f360-41f2-854a-4d0fb437af7d-etcd-client\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.610052 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.610403 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a6e6cf79-4d28-4262-801f-f85de17c5def-serving-cert\") pod \"console-operator-58897d9998-sdtpq\" (UID: \"a6e6cf79-4d28-4262-801f-f85de17c5def\") " pod="openshift-console-operator/console-operator-58897d9998-sdtpq" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.611379 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.612573 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.615539 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15bb3d2b-b085-4cf9-8ea1-1a67a5bc74e1-config\") pod \"openshift-apiserver-operator-796bbdcf4f-mm4gf\" (UID: \"15bb3d2b-b085-4cf9-8ea1-1a67a5bc74e1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mm4gf" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.631727 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.671335 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.691041 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.711000 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.731546 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.752324 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.771947 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.791484 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.811327 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.852966 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb5fn\" (UniqueName: \"kubernetes.io/projected/9a054bc8-9c59-47f8-8b37-7eb59a5331a0-kube-api-access-zb5fn\") pod \"machine-api-operator-5694c8668f-c7dks\" (UID: \"9a054bc8-9c59-47f8-8b37-7eb59a5331a0\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c7dks" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.875295 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.883158 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kptnx\" (UniqueName: \"kubernetes.io/projected/48be4455-d57d-412d-9f0d-191b038bcbcb-kube-api-access-kptnx\") pod \"cluster-samples-operator-665b6dd947-r2t4r\" (UID: \"48be4455-d57d-412d-9f0d-191b038bcbcb\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r2t4r" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.891532 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.911105 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.931498 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.951237 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.959303 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r2t4r" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.973394 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.993714 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 11 13:24:23 crc kubenswrapper[4923]: I0311 13:24:23.999458 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-c7dks" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.012112 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.021831 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.021995 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.022068 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.042968 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.051496 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.071909 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.118926 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2w9b\" (UniqueName: \"kubernetes.io/projected/e9b57443-86b7-45e7-a4f7-396719edfc4a-kube-api-access-w2w9b\") pod \"authentication-operator-69f744f599-4nhkq\" (UID: \"e9b57443-86b7-45e7-a4f7-396719edfc4a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-4nhkq" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.140738 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f4kq\" (UniqueName: \"kubernetes.io/projected/c7215b86-4f1e-44fb-b0a9-c7d261ce4a54-kube-api-access-2f4kq\") pod \"downloads-7954f5f757-b5jmc\" (UID: \"c7215b86-4f1e-44fb-b0a9-c7d261ce4a54\") " pod="openshift-console/downloads-7954f5f757-b5jmc" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.158248 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r582n\" (UniqueName: \"kubernetes.io/projected/16d21cac-dc41-43fb-a391-8864085617ed-kube-api-access-r582n\") pod \"controller-manager-879f6c89f-g25ck\" (UID: \"16d21cac-dc41-43fb-a391-8864085617ed\") " pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.171612 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.172765 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qgkl\" (UniqueName: \"kubernetes.io/projected/34272d37-e3bb-4315-b555-d1b4439657da-kube-api-access-6qgkl\") pod \"dns-operator-744455d44c-ndw7p\" (UID: \"34272d37-e3bb-4315-b555-d1b4439657da\") " pod="openshift-dns-operator/dns-operator-744455d44c-ndw7p" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.193179 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.194679 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-b5jmc" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.211814 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.231130 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.238306 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r2t4r"] Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.251173 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.270996 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.271214 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.277289 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-c7dks"] Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.288186 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-ndw7p" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.290256 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 11 13:24:24 crc kubenswrapper[4923]: W0311 13:24:24.301970 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a054bc8_9c59_47f8_8b37_7eb59a5331a0.slice/crio-590b970a09d8856120962febb6bce3b3fa642f5b81fda88ee6348ed6439802bc WatchSource:0}: Error finding container 590b970a09d8856120962febb6bce3b3fa642f5b81fda88ee6348ed6439802bc: Status 404 returned error can't find the container with id 590b970a09d8856120962febb6bce3b3fa642f5b81fda88ee6348ed6439802bc Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.311302 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.333060 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-4nhkq" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.333775 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.352847 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.372330 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.391733 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.406724 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-b5jmc"] Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.411777 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 11 13:24:24 crc kubenswrapper[4923]: W0311 13:24:24.421051 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7215b86_4f1e_44fb_b0a9_c7d261ce4a54.slice/crio-dff14b4270aac9ba6f5fb8944793179215b514556916d76f710992dc12e190aa WatchSource:0}: Error finding container dff14b4270aac9ba6f5fb8944793179215b514556916d76f710992dc12e190aa: Status 404 returned error can't find the container with id dff14b4270aac9ba6f5fb8944793179215b514556916d76f710992dc12e190aa Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.433279 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.451696 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.472529 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.489055 4923 request.go:700] Waited for 1.000396844s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-multus/secrets?fieldSelector=metadata.name%3Dmultus-admission-controller-secret&limit=500&resourceVersion=0 Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.491728 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.512701 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-g25ck"] Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.512747 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ndw7p"] Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.515191 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 11 13:24:24 crc kubenswrapper[4923]: W0311 13:24:24.522378 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16d21cac_dc41_43fb_a391_8864085617ed.slice/crio-f339e8b021e35c253199ac83505e1957927c26e46bb1b6044f6e75b9233de1ae WatchSource:0}: Error finding container f339e8b021e35c253199ac83505e1957927c26e46bb1b6044f6e75b9233de1ae: Status 404 returned error can't find the container with id f339e8b021e35c253199ac83505e1957927c26e46bb1b6044f6e75b9233de1ae Mar 11 13:24:24 crc kubenswrapper[4923]: W0311 13:24:24.527655 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34272d37_e3bb_4315_b555_d1b4439657da.slice/crio-da1a5ce56d6dcc8331ddfbdd26f8e549fdc583bf94bacf2a47f3e20e470b18eb WatchSource:0}: Error finding container da1a5ce56d6dcc8331ddfbdd26f8e549fdc583bf94bacf2a47f3e20e470b18eb: Status 404 returned error can't find the container with id da1a5ce56d6dcc8331ddfbdd26f8e549fdc583bf94bacf2a47f3e20e470b18eb Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.531654 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.551451 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.564217 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-4nhkq"] Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.572266 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 11 13:24:24 crc kubenswrapper[4923]: W0311 13:24:24.572900 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode9b57443_86b7_45e7_a4f7_396719edfc4a.slice/crio-76f1d4da1a07e07d8b22697c61941c1450bce63ef2df245ca01f2808839985b5 WatchSource:0}: Error finding container 76f1d4da1a07e07d8b22697c61941c1450bce63ef2df245ca01f2808839985b5: Status 404 returned error can't find the container with id 76f1d4da1a07e07d8b22697c61941c1450bce63ef2df245ca01f2808839985b5 Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.590493 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.611031 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.621379 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-b5jmc" event={"ID":"c7215b86-4f1e-44fb-b0a9-c7d261ce4a54","Type":"ContainerStarted","Data":"dff14b4270aac9ba6f5fb8944793179215b514556916d76f710992dc12e190aa"} Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.630607 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" event={"ID":"16d21cac-dc41-43fb-a391-8864085617ed","Type":"ContainerStarted","Data":"f339e8b021e35c253199ac83505e1957927c26e46bb1b6044f6e75b9233de1ae"} Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.631310 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.631870 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-4nhkq" event={"ID":"e9b57443-86b7-45e7-a4f7-396719edfc4a","Type":"ContainerStarted","Data":"76f1d4da1a07e07d8b22697c61941c1450bce63ef2df245ca01f2808839985b5"} Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.633430 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-c7dks" event={"ID":"9a054bc8-9c59-47f8-8b37-7eb59a5331a0","Type":"ContainerStarted","Data":"1d8d334b0823c61ef463a06b5443afa5b9fe3d456df08872c933e7a02d82da92"} Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.633477 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-c7dks" event={"ID":"9a054bc8-9c59-47f8-8b37-7eb59a5331a0","Type":"ContainerStarted","Data":"590b970a09d8856120962febb6bce3b3fa642f5b81fda88ee6348ed6439802bc"} Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.640240 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r2t4r" event={"ID":"48be4455-d57d-412d-9f0d-191b038bcbcb","Type":"ContainerStarted","Data":"996cb3b3120cb1428f2c15330ecb37833607008d9b2d1857d3f03e5756ebbbba"} Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.640281 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r2t4r" event={"ID":"48be4455-d57d-412d-9f0d-191b038bcbcb","Type":"ContainerStarted","Data":"d184076c11b4ab7f42e47c470c3aa45c4f997fc73cc30fcfa2dbb30a1bd74bb1"} Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.644510 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ndw7p" event={"ID":"34272d37-e3bb-4315-b555-d1b4439657da","Type":"ContainerStarted","Data":"da1a5ce56d6dcc8331ddfbdd26f8e549fdc583bf94bacf2a47f3e20e470b18eb"} Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.652614 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.674744 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.693146 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.711363 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.731679 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.751299 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.772481 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.791522 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.831970 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.851234 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.870990 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.891043 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.910919 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.931657 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.952142 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.971687 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 11 13:24:24 crc kubenswrapper[4923]: I0311 13:24:24.991429 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.012237 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.021056 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4nx54" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.031872 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.051467 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.072129 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.090789 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.110546 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.131098 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.157133 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.171152 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.191078 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.211724 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.232038 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.250640 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-sysctl-allowlist" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.271421 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.290541 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.310958 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.330864 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.351996 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.371327 4923 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.393003 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.415587 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.432961 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.452288 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.491198 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmsbn\" (UniqueName: \"kubernetes.io/projected/a6e6cf79-4d28-4262-801f-f85de17c5def-kube-api-access-pmsbn\") pod \"console-operator-58897d9998-sdtpq\" (UID: \"a6e6cf79-4d28-4262-801f-f85de17c5def\") " pod="openshift-console-operator/console-operator-58897d9998-sdtpq" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.509114 4923 request.go:700] Waited for 1.911763477s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver-operator/serviceaccounts/openshift-apiserver-operator/token Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.513469 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcgv5\" (UniqueName: \"kubernetes.io/projected/cfda8f36-49b9-4081-bbea-f311dd0a87a1-kube-api-access-mcgv5\") pod \"route-controller-manager-6576b87f9c-qk4ql\" (UID: \"cfda8f36-49b9-4081-bbea-f311dd0a87a1\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.526015 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lb7fp\" (UniqueName: \"kubernetes.io/projected/15bb3d2b-b085-4cf9-8ea1-1a67a5bc74e1-kube-api-access-lb7fp\") pod \"openshift-apiserver-operator-796bbdcf4f-mm4gf\" (UID: \"15bb3d2b-b085-4cf9-8ea1-1a67a5bc74e1\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mm4gf" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.544406 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm2p2\" (UniqueName: \"kubernetes.io/projected/f4826a3f-d90e-4b7c-a10b-864b78115463-kube-api-access-gm2p2\") pod \"cluster-image-registry-operator-dc59b4c8b-8bhpz\" (UID: \"f4826a3f-d90e-4b7c-a10b-864b78115463\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8bhpz" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.566133 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvrdv\" (UniqueName: \"kubernetes.io/projected/a67361c9-12cb-4a1a-893c-a735a58da9cd-kube-api-access-zvrdv\") pod \"apiserver-7bbb656c7d-rqlvv\" (UID: \"a67361c9-12cb-4a1a-893c-a735a58da9cd\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.569227 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.585581 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbvxd\" (UniqueName: \"kubernetes.io/projected/13238a87-3b9a-471a-8337-7e7f15018221-kube-api-access-xbvxd\") pod \"console-f9d7485db-mgrcg\" (UID: \"13238a87-3b9a-471a-8337-7e7f15018221\") " pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.608398 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jknjt\" (UniqueName: \"kubernetes.io/projected/919663b1-49e2-4405-b62a-5cd8b7092953-kube-api-access-jknjt\") pod \"openshift-config-operator-7777fb866f-l6bf9\" (UID: \"919663b1-49e2-4405-b62a-5cd8b7092953\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-l6bf9" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.627295 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-sdtpq" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.631741 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vd2z\" (UniqueName: \"kubernetes.io/projected/7c496315-8928-482f-be56-265af63f3888-kube-api-access-7vd2z\") pod \"etcd-operator-b45778765-7rzzf\" (UID: \"7c496315-8928-482f-be56-265af63f3888\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.632788 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.642991 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.645749 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm9c2\" (UniqueName: \"kubernetes.io/projected/5ec8d4f8-79c7-4153-94f1-0efe16a3da0f-kube-api-access-wm9c2\") pod \"machine-approver-56656f9798-jfhp5\" (UID: \"5ec8d4f8-79c7-4153-94f1-0efe16a3da0f\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jfhp5" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.648191 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" event={"ID":"16d21cac-dc41-43fb-a391-8864085617ed","Type":"ContainerStarted","Data":"bff21e897d1aa1ece817f88e76ea011624e18cc06b8cb1806c01ebcf4ddcb28e"} Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.648852 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.648948 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l6bf9" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.650397 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-4nhkq" event={"ID":"e9b57443-86b7-45e7-a4f7-396719edfc4a","Type":"ContainerStarted","Data":"41b67fe700aaa9bb298f5eb5fd14c02f7f5781c582329c42c07f0f96ffb531aa"} Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.651128 4923 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-g25ck container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.651161 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" podUID="16d21cac-dc41-43fb-a391-8864085617ed" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.658497 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-c7dks" event={"ID":"9a054bc8-9c59-47f8-8b37-7eb59a5331a0","Type":"ContainerStarted","Data":"a07989cb45d7a06582c8fc56b03d7a0f4dd2106db153ff3ef51ee26af8a7b0b0"} Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.662428 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.664933 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r2t4r" event={"ID":"48be4455-d57d-412d-9f0d-191b038bcbcb","Type":"ContainerStarted","Data":"1054a0aa42e2d2a0c75115510ef3480b74295e0929629a8a8da88249dd418e38"} Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.671043 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ndw7p" event={"ID":"34272d37-e3bb-4315-b555-d1b4439657da","Type":"ContainerStarted","Data":"e4f3893a77711f4ac86f00608e3f87ab6b55d292059cb91965edd4ce5dd11245"} Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.671086 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ndw7p" event={"ID":"34272d37-e3bb-4315-b555-d1b4439657da","Type":"ContainerStarted","Data":"c2cc9e19650b6d0d3c96a93a3ed774b08dbdcaefb6c89d64317307b90a444630"} Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.674534 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mm4gf" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.679662 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx5b4\" (UniqueName: \"kubernetes.io/projected/08068104-34d8-440d-8f98-f9a6216af858-kube-api-access-nx5b4\") pod \"oauth-openshift-558db77b4-hgkc2\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.680674 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-b5jmc" event={"ID":"c7215b86-4f1e-44fb-b0a9-c7d261ce4a54","Type":"ContainerStarted","Data":"588867230e1172988fc504bb2a5909c070282b5c9ed5090b5e7a0c01049692e4"} Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.680935 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-b5jmc" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.685528 4923 patch_prober.go:28] interesting pod/downloads-7954f5f757-b5jmc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" start-of-body= Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.685598 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b5jmc" podUID="c7215b86-4f1e-44fb-b0a9-c7d261ce4a54" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.690734 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f4826a3f-d90e-4b7c-a10b-864b78115463-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-8bhpz\" (UID: \"f4826a3f-d90e-4b7c-a10b-864b78115463\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8bhpz" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.705078 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbktl\" (UniqueName: \"kubernetes.io/projected/4890044b-8d18-4123-b029-7e2809294be1-kube-api-access-qbktl\") pod \"openshift-controller-manager-operator-756b6f6bc6-pk9tb\" (UID: \"4890044b-8d18-4123-b029-7e2809294be1\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pk9tb" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.726540 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h6lr\" (UniqueName: \"kubernetes.io/projected/1515573d-f360-41f2-854a-4d0fb437af7d-kube-api-access-8h6lr\") pod \"apiserver-76f77b778f-k6dxb\" (UID: \"1515573d-f360-41f2-854a-4d0fb437af7d\") " pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.751305 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.771403 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.792001 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.811201 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.814497 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql"] Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.822905 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.823059 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.823111 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.823151 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:25 crc kubenswrapper[4923]: E0311 13:24:25.823674 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:41.823645343 +0000 UTC m=+129.505547057 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.824647 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.827079 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.834460 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.834722 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8bhpz" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.851428 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.856929 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.869377 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.871742 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.902829 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jfhp5" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.910477 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.922235 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.923884 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/81c57b2e-1df9-44c1-9749-21fe5780f479-webhook-cert\") pod \"packageserver-d55dfcdfc-lc5r4\" (UID: \"81c57b2e-1df9-44c1-9749-21fe5780f479\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.923908 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3f8730fb-a1a2-440e-8fb2-1bf9134054e2-default-certificate\") pod \"router-default-5444994796-zd9ns\" (UID: \"3f8730fb-a1a2-440e-8fb2-1bf9134054e2\") " pod="openshift-ingress/router-default-5444994796-zd9ns" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.923925 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6b266fe-b225-4a79-a334-6be2b001c1b0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6v8k4\" (UID: \"c6b266fe-b225-4a79-a334-6be2b001c1b0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6v8k4" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.923950 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/714e692b-982f-4ba9-9d26-8819c83b14fe-config-volume\") pod \"dns-default-7lx27\" (UID: \"714e692b-982f-4ba9-9d26-8819c83b14fe\") " pod="openshift-dns/dns-default-7lx27" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.923989 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f8730fb-a1a2-440e-8fb2-1bf9134054e2-service-ca-bundle\") pod \"router-default-5444994796-zd9ns\" (UID: \"3f8730fb-a1a2-440e-8fb2-1bf9134054e2\") " pod="openshift-ingress/router-default-5444994796-zd9ns" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.924033 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2263dbb9-5bfb-4557-b1f2-2f33ef57db55-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-g96sx\" (UID: \"2263dbb9-5bfb-4557-b1f2-2f33ef57db55\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g96sx" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.924063 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3f8730fb-a1a2-440e-8fb2-1bf9134054e2-stats-auth\") pod \"router-default-5444994796-zd9ns\" (UID: \"3f8730fb-a1a2-440e-8fb2-1bf9134054e2\") " pod="openshift-ingress/router-default-5444994796-zd9ns" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.924109 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b819b4e5-85bd-4a31-bc2e-a119da615675-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.924126 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a47192bb-5458-4644-ba17-c9873f34922d-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-9c5lc\" (UID: \"a47192bb-5458-4644-ba17-c9873f34922d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9c5lc" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.924160 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/122f383b-c9fa-4ab3-8347-e75d16a0189b-config-volume\") pod \"collect-profiles-29553915-5tqf9\" (UID: \"122f383b-c9fa-4ab3-8347-e75d16a0189b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553915-5tqf9" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.924179 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ecd0070-8225-4da9-a577-4382d935f74b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-qnljg\" (UID: \"4ecd0070-8225-4da9-a577-4382d935f74b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qnljg" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.924226 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r7tg\" (UniqueName: \"kubernetes.io/projected/122f383b-c9fa-4ab3-8347-e75d16a0189b-kube-api-access-4r7tg\") pod \"collect-profiles-29553915-5tqf9\" (UID: \"122f383b-c9fa-4ab3-8347-e75d16a0189b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553915-5tqf9" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.924247 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz7df\" (UniqueName: \"kubernetes.io/projected/7e0c6535-0c45-4d8d-9b97-332888787b98-kube-api-access-rz7df\") pod \"migrator-59844c95c7-h7jfj\" (UID: \"7e0c6535-0c45-4d8d-9b97-332888787b98\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h7jfj" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.924366 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/122f383b-c9fa-4ab3-8347-e75d16a0189b-secret-volume\") pod \"collect-profiles-29553915-5tqf9\" (UID: \"122f383b-c9fa-4ab3-8347-e75d16a0189b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553915-5tqf9" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.924381 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0bfac15f-c184-44be-8490-81dfe546f94f-metrics-tls\") pod \"ingress-operator-5b745b69d9-2c4h7\" (UID: \"0bfac15f-c184-44be-8490-81dfe546f94f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2c4h7" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.924441 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0bfac15f-c184-44be-8490-81dfe546f94f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2c4h7\" (UID: \"0bfac15f-c184-44be-8490-81dfe546f94f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2c4h7" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.924476 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swzrr\" (UniqueName: \"kubernetes.io/projected/a47192bb-5458-4644-ba17-c9873f34922d-kube-api-access-swzrr\") pod \"multus-admission-controller-857f4d67dd-9c5lc\" (UID: \"a47192bb-5458-4644-ba17-c9873f34922d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9c5lc" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.924509 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l42rv\" (UniqueName: \"kubernetes.io/projected/2263dbb9-5bfb-4557-b1f2-2f33ef57db55-kube-api-access-l42rv\") pod \"kube-storage-version-migrator-operator-b67b599dd-g96sx\" (UID: \"2263dbb9-5bfb-4557-b1f2-2f33ef57db55\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g96sx" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.924544 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gk2xq\" (UniqueName: \"kubernetes.io/projected/3f8730fb-a1a2-440e-8fb2-1bf9134054e2-kube-api-access-gk2xq\") pod \"router-default-5444994796-zd9ns\" (UID: \"3f8730fb-a1a2-440e-8fb2-1bf9134054e2\") " pod="openshift-ingress/router-default-5444994796-zd9ns" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.924561 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3f8730fb-a1a2-440e-8fb2-1bf9134054e2-metrics-certs\") pod \"router-default-5444994796-zd9ns\" (UID: \"3f8730fb-a1a2-440e-8fb2-1bf9134054e2\") " pod="openshift-ingress/router-default-5444994796-zd9ns" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.924575 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/81c57b2e-1df9-44c1-9749-21fe5780f479-tmpfs\") pod \"packageserver-d55dfcdfc-lc5r4\" (UID: \"81c57b2e-1df9-44c1-9749-21fe5780f479\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.924592 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b819b4e5-85bd-4a31-bc2e-a119da615675-bound-sa-token\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.924607 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ecd0070-8225-4da9-a577-4382d935f74b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-qnljg\" (UID: \"4ecd0070-8225-4da9-a577-4382d935f74b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qnljg" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.924622 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c6b266fe-b225-4a79-a334-6be2b001c1b0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6v8k4\" (UID: \"c6b266fe-b225-4a79-a334-6be2b001c1b0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6v8k4" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.924672 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7bc92e0-e42d-46f4-bc56-5123cb6d36f0-config\") pod \"kube-controller-manager-operator-78b949d7b-dqr7s\" (UID: \"a7bc92e0-e42d-46f4-bc56-5123cb6d36f0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqr7s" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.924893 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfpt2\" (UniqueName: \"kubernetes.io/projected/0bfac15f-c184-44be-8490-81dfe546f94f-kube-api-access-rfpt2\") pod \"ingress-operator-5b745b69d9-2c4h7\" (UID: \"0bfac15f-c184-44be-8490-81dfe546f94f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2c4h7" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.924971 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gppsf\" (UniqueName: \"kubernetes.io/projected/81c57b2e-1df9-44c1-9749-21fe5780f479-kube-api-access-gppsf\") pod \"packageserver-d55dfcdfc-lc5r4\" (UID: \"81c57b2e-1df9-44c1-9749-21fe5780f479\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.925036 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0bfac15f-c184-44be-8490-81dfe546f94f-trusted-ca\") pod \"ingress-operator-5b745b69d9-2c4h7\" (UID: \"0bfac15f-c184-44be-8490-81dfe546f94f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2c4h7" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.925062 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7bc92e0-e42d-46f4-bc56-5123cb6d36f0-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-dqr7s\" (UID: \"a7bc92e0-e42d-46f4-bc56-5123cb6d36f0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqr7s" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.925082 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kg2m\" (UniqueName: \"kubernetes.io/projected/714e692b-982f-4ba9-9d26-8819c83b14fe-kube-api-access-7kg2m\") pod \"dns-default-7lx27\" (UID: \"714e692b-982f-4ba9-9d26-8819c83b14fe\") " pod="openshift-dns/dns-default-7lx27" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.925099 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rr5hc\" (UniqueName: \"kubernetes.io/projected/4f20534b-c922-4867-9294-fffaddb02fbf-kube-api-access-rr5hc\") pod \"catalog-operator-68c6474976-9grj8\" (UID: \"4f20534b-c922-4867-9294-fffaddb02fbf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9grj8" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.925209 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b819b4e5-85bd-4a31-bc2e-a119da615675-registry-tls\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.925227 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7fe6d35a-13be-4143-9072-5631cd656278-signing-key\") pod \"service-ca-9c57cc56f-6jmqc\" (UID: \"7fe6d35a-13be-4143-9072-5631cd656278\") " pod="openshift-service-ca/service-ca-9c57cc56f-6jmqc" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.925255 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqbp5\" (UniqueName: \"kubernetes.io/projected/7fe6d35a-13be-4143-9072-5631cd656278-kube-api-access-kqbp5\") pod \"service-ca-9c57cc56f-6jmqc\" (UID: \"7fe6d35a-13be-4143-9072-5631cd656278\") " pod="openshift-service-ca/service-ca-9c57cc56f-6jmqc" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.925317 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k8rp\" (UniqueName: \"kubernetes.io/projected/b819b4e5-85bd-4a31-bc2e-a119da615675-kube-api-access-8k8rp\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.926096 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/714e692b-982f-4ba9-9d26-8819c83b14fe-metrics-tls\") pod \"dns-default-7lx27\" (UID: \"714e692b-982f-4ba9-9d26-8819c83b14fe\") " pod="openshift-dns/dns-default-7lx27" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.926131 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4f20534b-c922-4867-9294-fffaddb02fbf-profile-collector-cert\") pod \"catalog-operator-68c6474976-9grj8\" (UID: \"4f20534b-c922-4867-9294-fffaddb02fbf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9grj8" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.926149 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7fe6d35a-13be-4143-9072-5631cd656278-signing-cabundle\") pod \"service-ca-9c57cc56f-6jmqc\" (UID: \"7fe6d35a-13be-4143-9072-5631cd656278\") " pod="openshift-service-ca/service-ca-9c57cc56f-6jmqc" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.926168 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a7bc92e0-e42d-46f4-bc56-5123cb6d36f0-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-dqr7s\" (UID: \"a7bc92e0-e42d-46f4-bc56-5123cb6d36f0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqr7s" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.926202 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b819b4e5-85bd-4a31-bc2e-a119da615675-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.926248 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ecd0070-8225-4da9-a577-4382d935f74b-config\") pod \"kube-apiserver-operator-766d6c64bb-qnljg\" (UID: \"4ecd0070-8225-4da9-a577-4382d935f74b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qnljg" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.926275 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b819b4e5-85bd-4a31-bc2e-a119da615675-trusted-ca\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.926292 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6b266fe-b225-4a79-a334-6be2b001c1b0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6v8k4\" (UID: \"c6b266fe-b225-4a79-a334-6be2b001c1b0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6v8k4" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.926380 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.926435 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b819b4e5-85bd-4a31-bc2e-a119da615675-registry-certificates\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.926452 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/81c57b2e-1df9-44c1-9749-21fe5780f479-apiservice-cert\") pod \"packageserver-d55dfcdfc-lc5r4\" (UID: \"81c57b2e-1df9-44c1-9749-21fe5780f479\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.926675 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2263dbb9-5bfb-4557-b1f2-2f33ef57db55-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-g96sx\" (UID: \"2263dbb9-5bfb-4557-b1f2-2f33ef57db55\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g96sx" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.926703 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4f20534b-c922-4867-9294-fffaddb02fbf-srv-cert\") pod \"catalog-operator-68c6474976-9grj8\" (UID: \"4f20534b-c922-4867-9294-fffaddb02fbf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9grj8" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.927020 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:25 crc kubenswrapper[4923]: E0311 13:24:25.928287 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:26.428276289 +0000 UTC m=+114.110178003 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.932324 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 11 13:24:25 crc kubenswrapper[4923]: I0311 13:24:25.968030 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pk9tb" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.028437 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.028657 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b0a9d170-2b10-48f1-81ab-898a7df26740-csi-data-dir\") pod \"csi-hostpathplugin-lc9xf\" (UID: \"b0a9d170-2b10-48f1-81ab-898a7df26740\") " pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.028689 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3f8730fb-a1a2-440e-8fb2-1bf9134054e2-stats-auth\") pod \"router-default-5444994796-zd9ns\" (UID: \"3f8730fb-a1a2-440e-8fb2-1bf9134054e2\") " pod="openshift-ingress/router-default-5444994796-zd9ns" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.028708 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gfx4\" (UniqueName: \"kubernetes.io/projected/b6887dd3-ec66-4364-8eed-f63615c35024-kube-api-access-9gfx4\") pod \"marketplace-operator-79b997595-vznqg\" (UID: \"b6887dd3-ec66-4364-8eed-f63615c35024\") " pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.028729 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b819b4e5-85bd-4a31-bc2e-a119da615675-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.028749 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a47192bb-5458-4644-ba17-c9873f34922d-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-9c5lc\" (UID: \"a47192bb-5458-4644-ba17-c9873f34922d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9c5lc" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.028766 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/122f383b-c9fa-4ab3-8347-e75d16a0189b-config-volume\") pod \"collect-profiles-29553915-5tqf9\" (UID: \"122f383b-c9fa-4ab3-8347-e75d16a0189b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553915-5tqf9" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.028789 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ecd0070-8225-4da9-a577-4382d935f74b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-qnljg\" (UID: \"4ecd0070-8225-4da9-a577-4382d935f74b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qnljg" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.028807 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-57kdl\" (UID: \"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04\") " pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.028824 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8jkj\" (UniqueName: \"kubernetes.io/projected/5a3187cd-89a0-4c86-acd9-e325360446eb-kube-api-access-w8jkj\") pod \"machine-config-operator-74547568cd-qkhqm\" (UID: \"5a3187cd-89a0-4c86-acd9-e325360446eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qkhqm" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.028840 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz7df\" (UniqueName: \"kubernetes.io/projected/7e0c6535-0c45-4d8d-9b97-332888787b98-kube-api-access-rz7df\") pod \"migrator-59844c95c7-h7jfj\" (UID: \"7e0c6535-0c45-4d8d-9b97-332888787b98\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h7jfj" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.028855 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5a3187cd-89a0-4c86-acd9-e325360446eb-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qkhqm\" (UID: \"5a3187cd-89a0-4c86-acd9-e325360446eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qkhqm" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.028871 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r7tg\" (UniqueName: \"kubernetes.io/projected/122f383b-c9fa-4ab3-8347-e75d16a0189b-kube-api-access-4r7tg\") pod \"collect-profiles-29553915-5tqf9\" (UID: \"122f383b-c9fa-4ab3-8347-e75d16a0189b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553915-5tqf9" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.028889 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/72930709-eec8-4c50-a0da-175e7cbea561-proxy-tls\") pod \"machine-config-controller-84d6567774-lb2kf\" (UID: \"72930709-eec8-4c50-a0da-175e7cbea561\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lb2kf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.028904 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9dbbbebf-e8db-40e2-9d1e-897a7b309c3a-certs\") pod \"machine-config-server-m4jm6\" (UID: \"9dbbbebf-e8db-40e2-9d1e-897a7b309c3a\") " pod="openshift-machine-config-operator/machine-config-server-m4jm6" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.028917 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5a3187cd-89a0-4c86-acd9-e325360446eb-proxy-tls\") pod \"machine-config-operator-74547568cd-qkhqm\" (UID: \"5a3187cd-89a0-4c86-acd9-e325360446eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qkhqm" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.028937 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j77fn\" (UniqueName: \"kubernetes.io/projected/00adaa61-ef98-4a06-9f60-4bd990cce287-kube-api-access-j77fn\") pod \"control-plane-machine-set-operator-78cbb6b69f-9cjnl\" (UID: \"00adaa61-ef98-4a06-9f60-4bd990cce287\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9cjnl" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.028964 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/122f383b-c9fa-4ab3-8347-e75d16a0189b-secret-volume\") pod \"collect-profiles-29553915-5tqf9\" (UID: \"122f383b-c9fa-4ab3-8347-e75d16a0189b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553915-5tqf9" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.028977 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0bfac15f-c184-44be-8490-81dfe546f94f-metrics-tls\") pod \"ingress-operator-5b745b69d9-2c4h7\" (UID: \"0bfac15f-c184-44be-8490-81dfe546f94f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2c4h7" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.028995 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6887dd3-ec66-4364-8eed-f63615c35024-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vznqg\" (UID: \"b6887dd3-ec66-4364-8eed-f63615c35024\") " pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029013 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swzrr\" (UniqueName: \"kubernetes.io/projected/a47192bb-5458-4644-ba17-c9873f34922d-kube-api-access-swzrr\") pod \"multus-admission-controller-857f4d67dd-9c5lc\" (UID: \"a47192bb-5458-4644-ba17-c9873f34922d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9c5lc" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029028 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0bfac15f-c184-44be-8490-81dfe546f94f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2c4h7\" (UID: \"0bfac15f-c184-44be-8490-81dfe546f94f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2c4h7" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029055 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l42rv\" (UniqueName: \"kubernetes.io/projected/2263dbb9-5bfb-4557-b1f2-2f33ef57db55-kube-api-access-l42rv\") pod \"kube-storage-version-migrator-operator-b67b599dd-g96sx\" (UID: \"2263dbb9-5bfb-4557-b1f2-2f33ef57db55\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g96sx" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029071 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntg7g\" (UniqueName: \"kubernetes.io/projected/9dbbbebf-e8db-40e2-9d1e-897a7b309c3a-kube-api-access-ntg7g\") pod \"machine-config-server-m4jm6\" (UID: \"9dbbbebf-e8db-40e2-9d1e-897a7b309c3a\") " pod="openshift-machine-config-operator/machine-config-server-m4jm6" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029087 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rj5bs\" (UniqueName: \"kubernetes.io/projected/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-kube-api-access-rj5bs\") pod \"cni-sysctl-allowlist-ds-57kdl\" (UID: \"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04\") " pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029101 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/867afc1e-15ff-4922-a2d2-5c0db0d7a019-cert\") pod \"ingress-canary-wdt5g\" (UID: \"867afc1e-15ff-4922-a2d2-5c0db0d7a019\") " pod="openshift-ingress-canary/ingress-canary-wdt5g" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029122 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gk2xq\" (UniqueName: \"kubernetes.io/projected/3f8730fb-a1a2-440e-8fb2-1bf9134054e2-kube-api-access-gk2xq\") pod \"router-default-5444994796-zd9ns\" (UID: \"3f8730fb-a1a2-440e-8fb2-1bf9134054e2\") " pod="openshift-ingress/router-default-5444994796-zd9ns" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029138 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3f8730fb-a1a2-440e-8fb2-1bf9134054e2-metrics-certs\") pod \"router-default-5444994796-zd9ns\" (UID: \"3f8730fb-a1a2-440e-8fb2-1bf9134054e2\") " pod="openshift-ingress/router-default-5444994796-zd9ns" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029154 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/81c57b2e-1df9-44c1-9749-21fe5780f479-tmpfs\") pod \"packageserver-d55dfcdfc-lc5r4\" (UID: \"81c57b2e-1df9-44c1-9749-21fe5780f479\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029177 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b819b4e5-85bd-4a31-bc2e-a119da615675-bound-sa-token\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029192 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh58q\" (UniqueName: \"kubernetes.io/projected/867afc1e-15ff-4922-a2d2-5c0db0d7a019-kube-api-access-nh58q\") pod \"ingress-canary-wdt5g\" (UID: \"867afc1e-15ff-4922-a2d2-5c0db0d7a019\") " pod="openshift-ingress-canary/ingress-canary-wdt5g" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029208 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ecd0070-8225-4da9-a577-4382d935f74b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-qnljg\" (UID: \"4ecd0070-8225-4da9-a577-4382d935f74b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qnljg" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029226 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c6b266fe-b225-4a79-a334-6be2b001c1b0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6v8k4\" (UID: \"c6b266fe-b225-4a79-a334-6be2b001c1b0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6v8k4" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029243 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9d5fee7-2298-4320-a8d1-c1bcb17c4917-config\") pod \"service-ca-operator-777779d784-dxlk5\" (UID: \"a9d5fee7-2298-4320-a8d1-c1bcb17c4917\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dxlk5" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029260 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7vh5\" (UniqueName: \"kubernetes.io/projected/717728c1-d998-4fc7-a0d2-ae90e3da4d43-kube-api-access-s7vh5\") pod \"package-server-manager-789f6589d5-h59nb\" (UID: \"717728c1-d998-4fc7-a0d2-ae90e3da4d43\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h59nb" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029276 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b407ce8d-2ba4-4744-a7d3-397efcc2e0b3-srv-cert\") pod \"olm-operator-6b444d44fb-5r4gt\" (UID: \"b407ce8d-2ba4-4744-a7d3-397efcc2e0b3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5r4gt" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029293 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7bc92e0-e42d-46f4-bc56-5123cb6d36f0-config\") pod \"kube-controller-manager-operator-78b949d7b-dqr7s\" (UID: \"a7bc92e0-e42d-46f4-bc56-5123cb6d36f0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqr7s" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029309 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rl5p\" (UniqueName: \"kubernetes.io/projected/b0a9d170-2b10-48f1-81ab-898a7df26740-kube-api-access-8rl5p\") pod \"csi-hostpathplugin-lc9xf\" (UID: \"b0a9d170-2b10-48f1-81ab-898a7df26740\") " pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029326 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfpt2\" (UniqueName: \"kubernetes.io/projected/0bfac15f-c184-44be-8490-81dfe546f94f-kube-api-access-rfpt2\") pod \"ingress-operator-5b745b69d9-2c4h7\" (UID: \"0bfac15f-c184-44be-8490-81dfe546f94f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2c4h7" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029367 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6887dd3-ec66-4364-8eed-f63615c35024-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vznqg\" (UID: \"b6887dd3-ec66-4364-8eed-f63615c35024\") " pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029384 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gppsf\" (UniqueName: \"kubernetes.io/projected/81c57b2e-1df9-44c1-9749-21fe5780f479-kube-api-access-gppsf\") pod \"packageserver-d55dfcdfc-lc5r4\" (UID: \"81c57b2e-1df9-44c1-9749-21fe5780f479\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029401 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b0a9d170-2b10-48f1-81ab-898a7df26740-plugins-dir\") pod \"csi-hostpathplugin-lc9xf\" (UID: \"b0a9d170-2b10-48f1-81ab-898a7df26740\") " pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029416 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/72930709-eec8-4c50-a0da-175e7cbea561-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-lb2kf\" (UID: \"72930709-eec8-4c50-a0da-175e7cbea561\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lb2kf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029438 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/717728c1-d998-4fc7-a0d2-ae90e3da4d43-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-h59nb\" (UID: \"717728c1-d998-4fc7-a0d2-ae90e3da4d43\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h59nb" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029455 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0bfac15f-c184-44be-8490-81dfe546f94f-trusted-ca\") pod \"ingress-operator-5b745b69d9-2c4h7\" (UID: \"0bfac15f-c184-44be-8490-81dfe546f94f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2c4h7" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029470 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7bc92e0-e42d-46f4-bc56-5123cb6d36f0-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-dqr7s\" (UID: \"a7bc92e0-e42d-46f4-bc56-5123cb6d36f0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqr7s" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029488 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kg2m\" (UniqueName: \"kubernetes.io/projected/714e692b-982f-4ba9-9d26-8819c83b14fe-kube-api-access-7kg2m\") pod \"dns-default-7lx27\" (UID: \"714e692b-982f-4ba9-9d26-8819c83b14fe\") " pod="openshift-dns/dns-default-7lx27" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029504 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rr5hc\" (UniqueName: \"kubernetes.io/projected/4f20534b-c922-4867-9294-fffaddb02fbf-kube-api-access-rr5hc\") pod \"catalog-operator-68c6474976-9grj8\" (UID: \"4f20534b-c922-4867-9294-fffaddb02fbf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9grj8" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029521 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b0a9d170-2b10-48f1-81ab-898a7df26740-registration-dir\") pod \"csi-hostpathplugin-lc9xf\" (UID: \"b0a9d170-2b10-48f1-81ab-898a7df26740\") " pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029537 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6x696\" (UniqueName: \"kubernetes.io/projected/a9d5fee7-2298-4320-a8d1-c1bcb17c4917-kube-api-access-6x696\") pod \"service-ca-operator-777779d784-dxlk5\" (UID: \"a9d5fee7-2298-4320-a8d1-c1bcb17c4917\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dxlk5" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029553 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b819b4e5-85bd-4a31-bc2e-a119da615675-registry-tls\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029569 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7fe6d35a-13be-4143-9072-5631cd656278-signing-key\") pod \"service-ca-9c57cc56f-6jmqc\" (UID: \"7fe6d35a-13be-4143-9072-5631cd656278\") " pod="openshift-service-ca/service-ca-9c57cc56f-6jmqc" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029584 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b0a9d170-2b10-48f1-81ab-898a7df26740-mountpoint-dir\") pod \"csi-hostpathplugin-lc9xf\" (UID: \"b0a9d170-2b10-48f1-81ab-898a7df26740\") " pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029611 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b0a9d170-2b10-48f1-81ab-898a7df26740-socket-dir\") pod \"csi-hostpathplugin-lc9xf\" (UID: \"b0a9d170-2b10-48f1-81ab-898a7df26740\") " pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029628 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4f20534b-c922-4867-9294-fffaddb02fbf-profile-collector-cert\") pod \"catalog-operator-68c6474976-9grj8\" (UID: \"4f20534b-c922-4867-9294-fffaddb02fbf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9grj8" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029645 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7fe6d35a-13be-4143-9072-5631cd656278-signing-cabundle\") pod \"service-ca-9c57cc56f-6jmqc\" (UID: \"7fe6d35a-13be-4143-9072-5631cd656278\") " pod="openshift-service-ca/service-ca-9c57cc56f-6jmqc" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029662 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqbp5\" (UniqueName: \"kubernetes.io/projected/7fe6d35a-13be-4143-9072-5631cd656278-kube-api-access-kqbp5\") pod \"service-ca-9c57cc56f-6jmqc\" (UID: \"7fe6d35a-13be-4143-9072-5631cd656278\") " pod="openshift-service-ca/service-ca-9c57cc56f-6jmqc" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029680 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9d5fee7-2298-4320-a8d1-c1bcb17c4917-serving-cert\") pod \"service-ca-operator-777779d784-dxlk5\" (UID: \"a9d5fee7-2298-4320-a8d1-c1bcb17c4917\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dxlk5" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029700 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k8rp\" (UniqueName: \"kubernetes.io/projected/b819b4e5-85bd-4a31-bc2e-a119da615675-kube-api-access-8k8rp\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029715 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/714e692b-982f-4ba9-9d26-8819c83b14fe-metrics-tls\") pod \"dns-default-7lx27\" (UID: \"714e692b-982f-4ba9-9d26-8819c83b14fe\") " pod="openshift-dns/dns-default-7lx27" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029731 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a7bc92e0-e42d-46f4-bc56-5123cb6d36f0-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-dqr7s\" (UID: \"a7bc92e0-e42d-46f4-bc56-5123cb6d36f0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqr7s" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029747 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b819b4e5-85bd-4a31-bc2e-a119da615675-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029775 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ecd0070-8225-4da9-a577-4382d935f74b-config\") pod \"kube-apiserver-operator-766d6c64bb-qnljg\" (UID: \"4ecd0070-8225-4da9-a577-4382d935f74b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qnljg" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029792 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px8lm\" (UniqueName: \"kubernetes.io/projected/b407ce8d-2ba4-4744-a7d3-397efcc2e0b3-kube-api-access-px8lm\") pod \"olm-operator-6b444d44fb-5r4gt\" (UID: \"b407ce8d-2ba4-4744-a7d3-397efcc2e0b3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5r4gt" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029810 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6b266fe-b225-4a79-a334-6be2b001c1b0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6v8k4\" (UID: \"c6b266fe-b225-4a79-a334-6be2b001c1b0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6v8k4" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029827 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b407ce8d-2ba4-4744-a7d3-397efcc2e0b3-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5r4gt\" (UID: \"b407ce8d-2ba4-4744-a7d3-397efcc2e0b3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5r4gt" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029852 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b819b4e5-85bd-4a31-bc2e-a119da615675-trusted-ca\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029870 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/00adaa61-ef98-4a06-9f60-4bd990cce287-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9cjnl\" (UID: \"00adaa61-ef98-4a06-9f60-4bd990cce287\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9cjnl" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029888 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b819b4e5-85bd-4a31-bc2e-a119da615675-registry-certificates\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029903 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/81c57b2e-1df9-44c1-9749-21fe5780f479-apiservice-cert\") pod \"packageserver-d55dfcdfc-lc5r4\" (UID: \"81c57b2e-1df9-44c1-9749-21fe5780f479\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029920 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4f20534b-c922-4867-9294-fffaddb02fbf-srv-cert\") pod \"catalog-operator-68c6474976-9grj8\" (UID: \"4f20534b-c922-4867-9294-fffaddb02fbf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9grj8" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029939 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2263dbb9-5bfb-4557-b1f2-2f33ef57db55-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-g96sx\" (UID: \"2263dbb9-5bfb-4557-b1f2-2f33ef57db55\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g96sx" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029966 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-ready\") pod \"cni-sysctl-allowlist-ds-57kdl\" (UID: \"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04\") " pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.029988 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9dbbbebf-e8db-40e2-9d1e-897a7b309c3a-node-bootstrap-token\") pod \"machine-config-server-m4jm6\" (UID: \"9dbbbebf-e8db-40e2-9d1e-897a7b309c3a\") " pod="openshift-machine-config-operator/machine-config-server-m4jm6" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.030004 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5a3187cd-89a0-4c86-acd9-e325360446eb-images\") pod \"machine-config-operator-74547568cd-qkhqm\" (UID: \"5a3187cd-89a0-4c86-acd9-e325360446eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qkhqm" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.030021 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/81c57b2e-1df9-44c1-9749-21fe5780f479-webhook-cert\") pod \"packageserver-d55dfcdfc-lc5r4\" (UID: \"81c57b2e-1df9-44c1-9749-21fe5780f479\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.030046 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3f8730fb-a1a2-440e-8fb2-1bf9134054e2-default-certificate\") pod \"router-default-5444994796-zd9ns\" (UID: \"3f8730fb-a1a2-440e-8fb2-1bf9134054e2\") " pod="openshift-ingress/router-default-5444994796-zd9ns" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.030062 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6b266fe-b225-4a79-a334-6be2b001c1b0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6v8k4\" (UID: \"c6b266fe-b225-4a79-a334-6be2b001c1b0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6v8k4" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.030077 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58rk8\" (UniqueName: \"kubernetes.io/projected/72930709-eec8-4c50-a0da-175e7cbea561-kube-api-access-58rk8\") pod \"machine-config-controller-84d6567774-lb2kf\" (UID: \"72930709-eec8-4c50-a0da-175e7cbea561\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lb2kf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.030092 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/714e692b-982f-4ba9-9d26-8819c83b14fe-config-volume\") pod \"dns-default-7lx27\" (UID: \"714e692b-982f-4ba9-9d26-8819c83b14fe\") " pod="openshift-dns/dns-default-7lx27" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.030108 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-57kdl\" (UID: \"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04\") " pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.030124 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f8730fb-a1a2-440e-8fb2-1bf9134054e2-service-ca-bundle\") pod \"router-default-5444994796-zd9ns\" (UID: \"3f8730fb-a1a2-440e-8fb2-1bf9134054e2\") " pod="openshift-ingress/router-default-5444994796-zd9ns" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.030141 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2263dbb9-5bfb-4557-b1f2-2f33ef57db55-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-g96sx\" (UID: \"2263dbb9-5bfb-4557-b1f2-2f33ef57db55\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g96sx" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.034988 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2263dbb9-5bfb-4557-b1f2-2f33ef57db55-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-g96sx\" (UID: \"2263dbb9-5bfb-4557-b1f2-2f33ef57db55\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g96sx" Mar 11 13:24:26 crc kubenswrapper[4923]: E0311 13:24:26.035073 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:26.53505846 +0000 UTC m=+114.216960174 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.042050 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/122f383b-c9fa-4ab3-8347-e75d16a0189b-config-volume\") pod \"collect-profiles-29553915-5tqf9\" (UID: \"122f383b-c9fa-4ab3-8347-e75d16a0189b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553915-5tqf9" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.043088 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c6b266fe-b225-4a79-a334-6be2b001c1b0-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6v8k4\" (UID: \"c6b266fe-b225-4a79-a334-6be2b001c1b0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6v8k4" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.043330 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0bfac15f-c184-44be-8490-81dfe546f94f-trusted-ca\") pod \"ingress-operator-5b745b69d9-2c4h7\" (UID: \"0bfac15f-c184-44be-8490-81dfe546f94f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2c4h7" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.044249 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/3f8730fb-a1a2-440e-8fb2-1bf9134054e2-stats-auth\") pod \"router-default-5444994796-zd9ns\" (UID: \"3f8730fb-a1a2-440e-8fb2-1bf9134054e2\") " pod="openshift-ingress/router-default-5444994796-zd9ns" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.044794 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/81c57b2e-1df9-44c1-9749-21fe5780f479-tmpfs\") pod \"packageserver-d55dfcdfc-lc5r4\" (UID: \"81c57b2e-1df9-44c1-9749-21fe5780f479\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.044975 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/714e692b-982f-4ba9-9d26-8819c83b14fe-config-volume\") pod \"dns-default-7lx27\" (UID: \"714e692b-982f-4ba9-9d26-8819c83b14fe\") " pod="openshift-dns/dns-default-7lx27" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.045630 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3f8730fb-a1a2-440e-8fb2-1bf9134054e2-service-ca-bundle\") pod \"router-default-5444994796-zd9ns\" (UID: \"3f8730fb-a1a2-440e-8fb2-1bf9134054e2\") " pod="openshift-ingress/router-default-5444994796-zd9ns" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.046756 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b819b4e5-85bd-4a31-bc2e-a119da615675-registry-certificates\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.047640 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2263dbb9-5bfb-4557-b1f2-2f33ef57db55-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-g96sx\" (UID: \"2263dbb9-5bfb-4557-b1f2-2f33ef57db55\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g96sx" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.051754 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/7fe6d35a-13be-4143-9072-5631cd656278-signing-cabundle\") pod \"service-ca-9c57cc56f-6jmqc\" (UID: \"7fe6d35a-13be-4143-9072-5631cd656278\") " pod="openshift-service-ca/service-ca-9c57cc56f-6jmqc" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.052082 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4f20534b-c922-4867-9294-fffaddb02fbf-profile-collector-cert\") pod \"catalog-operator-68c6474976-9grj8\" (UID: \"4f20534b-c922-4867-9294-fffaddb02fbf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9grj8" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.054640 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/3f8730fb-a1a2-440e-8fb2-1bf9134054e2-default-certificate\") pod \"router-default-5444994796-zd9ns\" (UID: \"3f8730fb-a1a2-440e-8fb2-1bf9134054e2\") " pod="openshift-ingress/router-default-5444994796-zd9ns" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.055006 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b819b4e5-85bd-4a31-bc2e-a119da615675-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.055462 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/81c57b2e-1df9-44c1-9749-21fe5780f479-webhook-cert\") pod \"packageserver-d55dfcdfc-lc5r4\" (UID: \"81c57b2e-1df9-44c1-9749-21fe5780f479\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.056073 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b819b4e5-85bd-4a31-bc2e-a119da615675-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.056525 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7bc92e0-e42d-46f4-bc56-5123cb6d36f0-config\") pod \"kube-controller-manager-operator-78b949d7b-dqr7s\" (UID: \"a7bc92e0-e42d-46f4-bc56-5123cb6d36f0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqr7s" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.056861 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ecd0070-8225-4da9-a577-4382d935f74b-config\") pod \"kube-apiserver-operator-766d6c64bb-qnljg\" (UID: \"4ecd0070-8225-4da9-a577-4382d935f74b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qnljg" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.057106 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a47192bb-5458-4644-ba17-c9873f34922d-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-9c5lc\" (UID: \"a47192bb-5458-4644-ba17-c9873f34922d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9c5lc" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.061682 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/122f383b-c9fa-4ab3-8347-e75d16a0189b-secret-volume\") pod \"collect-profiles-29553915-5tqf9\" (UID: \"122f383b-c9fa-4ab3-8347-e75d16a0189b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553915-5tqf9" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.061714 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b819b4e5-85bd-4a31-bc2e-a119da615675-trusted-ca\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.062526 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/7fe6d35a-13be-4143-9072-5631cd656278-signing-key\") pod \"service-ca-9c57cc56f-6jmqc\" (UID: \"7fe6d35a-13be-4143-9072-5631cd656278\") " pod="openshift-service-ca/service-ca-9c57cc56f-6jmqc" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.063006 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/714e692b-982f-4ba9-9d26-8819c83b14fe-metrics-tls\") pod \"dns-default-7lx27\" (UID: \"714e692b-982f-4ba9-9d26-8819c83b14fe\") " pod="openshift-dns/dns-default-7lx27" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.072108 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0bfac15f-c184-44be-8490-81dfe546f94f-metrics-tls\") pod \"ingress-operator-5b745b69d9-2c4h7\" (UID: \"0bfac15f-c184-44be-8490-81dfe546f94f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2c4h7" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.087403 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz7df\" (UniqueName: \"kubernetes.io/projected/7e0c6535-0c45-4d8d-9b97-332888787b98-kube-api-access-rz7df\") pod \"migrator-59844c95c7-h7jfj\" (UID: \"7e0c6535-0c45-4d8d-9b97-332888787b98\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h7jfj" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.087890 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4ecd0070-8225-4da9-a577-4382d935f74b-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-qnljg\" (UID: \"4ecd0070-8225-4da9-a577-4382d935f74b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qnljg" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.088354 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ecd0070-8225-4da9-a577-4382d935f74b-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-qnljg\" (UID: \"4ecd0070-8225-4da9-a577-4382d935f74b\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qnljg" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.099255 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7bc92e0-e42d-46f4-bc56-5123cb6d36f0-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-dqr7s\" (UID: \"a7bc92e0-e42d-46f4-bc56-5123cb6d36f0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqr7s" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.102023 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/81c57b2e-1df9-44c1-9749-21fe5780f479-apiservice-cert\") pod \"packageserver-d55dfcdfc-lc5r4\" (UID: \"81c57b2e-1df9-44c1-9749-21fe5780f479\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.102023 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b819b4e5-85bd-4a31-bc2e-a119da615675-registry-tls\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.102030 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3f8730fb-a1a2-440e-8fb2-1bf9134054e2-metrics-certs\") pod \"router-default-5444994796-zd9ns\" (UID: \"3f8730fb-a1a2-440e-8fb2-1bf9134054e2\") " pod="openshift-ingress/router-default-5444994796-zd9ns" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.102389 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c6b266fe-b225-4a79-a334-6be2b001c1b0-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6v8k4\" (UID: \"c6b266fe-b225-4a79-a334-6be2b001c1b0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6v8k4" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.106767 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4f20534b-c922-4867-9294-fffaddb02fbf-srv-cert\") pod \"catalog-operator-68c6474976-9grj8\" (UID: \"4f20534b-c922-4867-9294-fffaddb02fbf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9grj8" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.117478 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-sdtpq"] Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.126120 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0bfac15f-c184-44be-8490-81dfe546f94f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-2c4h7\" (UID: \"0bfac15f-c184-44be-8490-81dfe546f94f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2c4h7" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.131944 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b407ce8d-2ba4-4744-a7d3-397efcc2e0b3-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5r4gt\" (UID: \"b407ce8d-2ba4-4744-a7d3-397efcc2e0b3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5r4gt" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.131977 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/00adaa61-ef98-4a06-9f60-4bd990cce287-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9cjnl\" (UID: \"00adaa61-ef98-4a06-9f60-4bd990cce287\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9cjnl" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132000 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-ready\") pod \"cni-sysctl-allowlist-ds-57kdl\" (UID: \"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04\") " pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132023 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132040 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9dbbbebf-e8db-40e2-9d1e-897a7b309c3a-node-bootstrap-token\") pod \"machine-config-server-m4jm6\" (UID: \"9dbbbebf-e8db-40e2-9d1e-897a7b309c3a\") " pod="openshift-machine-config-operator/machine-config-server-m4jm6" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132061 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5a3187cd-89a0-4c86-acd9-e325360446eb-images\") pod \"machine-config-operator-74547568cd-qkhqm\" (UID: \"5a3187cd-89a0-4c86-acd9-e325360446eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qkhqm" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132081 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58rk8\" (UniqueName: \"kubernetes.io/projected/72930709-eec8-4c50-a0da-175e7cbea561-kube-api-access-58rk8\") pod \"machine-config-controller-84d6567774-lb2kf\" (UID: \"72930709-eec8-4c50-a0da-175e7cbea561\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lb2kf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132099 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-57kdl\" (UID: \"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04\") " pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132116 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b0a9d170-2b10-48f1-81ab-898a7df26740-csi-data-dir\") pod \"csi-hostpathplugin-lc9xf\" (UID: \"b0a9d170-2b10-48f1-81ab-898a7df26740\") " pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132133 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gfx4\" (UniqueName: \"kubernetes.io/projected/b6887dd3-ec66-4364-8eed-f63615c35024-kube-api-access-9gfx4\") pod \"marketplace-operator-79b997595-vznqg\" (UID: \"b6887dd3-ec66-4364-8eed-f63615c35024\") " pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132150 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-57kdl\" (UID: \"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04\") " pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132165 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8jkj\" (UniqueName: \"kubernetes.io/projected/5a3187cd-89a0-4c86-acd9-e325360446eb-kube-api-access-w8jkj\") pod \"machine-config-operator-74547568cd-qkhqm\" (UID: \"5a3187cd-89a0-4c86-acd9-e325360446eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qkhqm" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132185 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5a3187cd-89a0-4c86-acd9-e325360446eb-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qkhqm\" (UID: \"5a3187cd-89a0-4c86-acd9-e325360446eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qkhqm" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132198 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/72930709-eec8-4c50-a0da-175e7cbea561-proxy-tls\") pod \"machine-config-controller-84d6567774-lb2kf\" (UID: \"72930709-eec8-4c50-a0da-175e7cbea561\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lb2kf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132213 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9dbbbebf-e8db-40e2-9d1e-897a7b309c3a-certs\") pod \"machine-config-server-m4jm6\" (UID: \"9dbbbebf-e8db-40e2-9d1e-897a7b309c3a\") " pod="openshift-machine-config-operator/machine-config-server-m4jm6" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132243 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5a3187cd-89a0-4c86-acd9-e325360446eb-proxy-tls\") pod \"machine-config-operator-74547568cd-qkhqm\" (UID: \"5a3187cd-89a0-4c86-acd9-e325360446eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qkhqm" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132260 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j77fn\" (UniqueName: \"kubernetes.io/projected/00adaa61-ef98-4a06-9f60-4bd990cce287-kube-api-access-j77fn\") pod \"control-plane-machine-set-operator-78cbb6b69f-9cjnl\" (UID: \"00adaa61-ef98-4a06-9f60-4bd990cce287\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9cjnl" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132279 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6887dd3-ec66-4364-8eed-f63615c35024-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vznqg\" (UID: \"b6887dd3-ec66-4364-8eed-f63615c35024\") " pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132308 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntg7g\" (UniqueName: \"kubernetes.io/projected/9dbbbebf-e8db-40e2-9d1e-897a7b309c3a-kube-api-access-ntg7g\") pod \"machine-config-server-m4jm6\" (UID: \"9dbbbebf-e8db-40e2-9d1e-897a7b309c3a\") " pod="openshift-machine-config-operator/machine-config-server-m4jm6" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132330 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rj5bs\" (UniqueName: \"kubernetes.io/projected/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-kube-api-access-rj5bs\") pod \"cni-sysctl-allowlist-ds-57kdl\" (UID: \"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04\") " pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132357 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/867afc1e-15ff-4922-a2d2-5c0db0d7a019-cert\") pod \"ingress-canary-wdt5g\" (UID: \"867afc1e-15ff-4922-a2d2-5c0db0d7a019\") " pod="openshift-ingress-canary/ingress-canary-wdt5g" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132378 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh58q\" (UniqueName: \"kubernetes.io/projected/867afc1e-15ff-4922-a2d2-5c0db0d7a019-kube-api-access-nh58q\") pod \"ingress-canary-wdt5g\" (UID: \"867afc1e-15ff-4922-a2d2-5c0db0d7a019\") " pod="openshift-ingress-canary/ingress-canary-wdt5g" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132399 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9d5fee7-2298-4320-a8d1-c1bcb17c4917-config\") pod \"service-ca-operator-777779d784-dxlk5\" (UID: \"a9d5fee7-2298-4320-a8d1-c1bcb17c4917\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dxlk5" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132419 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7vh5\" (UniqueName: \"kubernetes.io/projected/717728c1-d998-4fc7-a0d2-ae90e3da4d43-kube-api-access-s7vh5\") pod \"package-server-manager-789f6589d5-h59nb\" (UID: \"717728c1-d998-4fc7-a0d2-ae90e3da4d43\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h59nb" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132435 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rl5p\" (UniqueName: \"kubernetes.io/projected/b0a9d170-2b10-48f1-81ab-898a7df26740-kube-api-access-8rl5p\") pod \"csi-hostpathplugin-lc9xf\" (UID: \"b0a9d170-2b10-48f1-81ab-898a7df26740\") " pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132450 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b407ce8d-2ba4-4744-a7d3-397efcc2e0b3-srv-cert\") pod \"olm-operator-6b444d44fb-5r4gt\" (UID: \"b407ce8d-2ba4-4744-a7d3-397efcc2e0b3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5r4gt" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132480 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6887dd3-ec66-4364-8eed-f63615c35024-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vznqg\" (UID: \"b6887dd3-ec66-4364-8eed-f63615c35024\") " pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132497 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/72930709-eec8-4c50-a0da-175e7cbea561-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-lb2kf\" (UID: \"72930709-eec8-4c50-a0da-175e7cbea561\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lb2kf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132513 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b0a9d170-2b10-48f1-81ab-898a7df26740-plugins-dir\") pod \"csi-hostpathplugin-lc9xf\" (UID: \"b0a9d170-2b10-48f1-81ab-898a7df26740\") " pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132528 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/717728c1-d998-4fc7-a0d2-ae90e3da4d43-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-h59nb\" (UID: \"717728c1-d998-4fc7-a0d2-ae90e3da4d43\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h59nb" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132578 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b0a9d170-2b10-48f1-81ab-898a7df26740-registration-dir\") pod \"csi-hostpathplugin-lc9xf\" (UID: \"b0a9d170-2b10-48f1-81ab-898a7df26740\") " pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132600 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6x696\" (UniqueName: \"kubernetes.io/projected/a9d5fee7-2298-4320-a8d1-c1bcb17c4917-kube-api-access-6x696\") pod \"service-ca-operator-777779d784-dxlk5\" (UID: \"a9d5fee7-2298-4320-a8d1-c1bcb17c4917\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dxlk5" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132624 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b0a9d170-2b10-48f1-81ab-898a7df26740-mountpoint-dir\") pod \"csi-hostpathplugin-lc9xf\" (UID: \"b0a9d170-2b10-48f1-81ab-898a7df26740\") " pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132647 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b0a9d170-2b10-48f1-81ab-898a7df26740-socket-dir\") pod \"csi-hostpathplugin-lc9xf\" (UID: \"b0a9d170-2b10-48f1-81ab-898a7df26740\") " pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132682 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9d5fee7-2298-4320-a8d1-c1bcb17c4917-serving-cert\") pod \"service-ca-operator-777779d784-dxlk5\" (UID: \"a9d5fee7-2298-4320-a8d1-c1bcb17c4917\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dxlk5" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.132711 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px8lm\" (UniqueName: \"kubernetes.io/projected/b407ce8d-2ba4-4744-a7d3-397efcc2e0b3-kube-api-access-px8lm\") pod \"olm-operator-6b444d44fb-5r4gt\" (UID: \"b407ce8d-2ba4-4744-a7d3-397efcc2e0b3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5r4gt" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.134093 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6887dd3-ec66-4364-8eed-f63615c35024-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vznqg\" (UID: \"b6887dd3-ec66-4364-8eed-f63615c35024\") " pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.134715 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/b0a9d170-2b10-48f1-81ab-898a7df26740-csi-data-dir\") pod \"csi-hostpathplugin-lc9xf\" (UID: \"b0a9d170-2b10-48f1-81ab-898a7df26740\") " pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.135387 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5a3187cd-89a0-4c86-acd9-e325360446eb-auth-proxy-config\") pod \"machine-config-operator-74547568cd-qkhqm\" (UID: \"5a3187cd-89a0-4c86-acd9-e325360446eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qkhqm" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.135553 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a9d5fee7-2298-4320-a8d1-c1bcb17c4917-config\") pod \"service-ca-operator-777779d784-dxlk5\" (UID: \"a9d5fee7-2298-4320-a8d1-c1bcb17c4917\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dxlk5" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.135643 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/b0a9d170-2b10-48f1-81ab-898a7df26740-registration-dir\") pod \"csi-hostpathplugin-lc9xf\" (UID: \"b0a9d170-2b10-48f1-81ab-898a7df26740\") " pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.135671 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/b0a9d170-2b10-48f1-81ab-898a7df26740-plugins-dir\") pod \"csi-hostpathplugin-lc9xf\" (UID: \"b0a9d170-2b10-48f1-81ab-898a7df26740\") " pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.136024 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/b0a9d170-2b10-48f1-81ab-898a7df26740-mountpoint-dir\") pod \"csi-hostpathplugin-lc9xf\" (UID: \"b0a9d170-2b10-48f1-81ab-898a7df26740\") " pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.136071 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/b0a9d170-2b10-48f1-81ab-898a7df26740-socket-dir\") pod \"csi-hostpathplugin-lc9xf\" (UID: \"b0a9d170-2b10-48f1-81ab-898a7df26740\") " pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.137585 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-ready\") pod \"cni-sysctl-allowlist-ds-57kdl\" (UID: \"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04\") " pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.137790 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-57kdl\" (UID: \"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04\") " pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" Mar 11 13:24:26 crc kubenswrapper[4923]: E0311 13:24:26.138154 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:26.638138919 +0000 UTC m=+114.320040623 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.138619 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/5a3187cd-89a0-4c86-acd9-e325360446eb-images\") pod \"machine-config-operator-74547568cd-qkhqm\" (UID: \"5a3187cd-89a0-4c86-acd9-e325360446eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qkhqm" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.139367 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r7tg\" (UniqueName: \"kubernetes.io/projected/122f383b-c9fa-4ab3-8347-e75d16a0189b-kube-api-access-4r7tg\") pod \"collect-profiles-29553915-5tqf9\" (UID: \"122f383b-c9fa-4ab3-8347-e75d16a0189b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553915-5tqf9" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.139326 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6887dd3-ec66-4364-8eed-f63615c35024-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vznqg\" (UID: \"b6887dd3-ec66-4364-8eed-f63615c35024\") " pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.143046 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/72930709-eec8-4c50-a0da-175e7cbea561-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-lb2kf\" (UID: \"72930709-eec8-4c50-a0da-175e7cbea561\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lb2kf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.143492 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/9dbbbebf-e8db-40e2-9d1e-897a7b309c3a-node-bootstrap-token\") pod \"machine-config-server-m4jm6\" (UID: \"9dbbbebf-e8db-40e2-9d1e-897a7b309c3a\") " pod="openshift-machine-config-operator/machine-config-server-m4jm6" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.143794 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b407ce8d-2ba4-4744-a7d3-397efcc2e0b3-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5r4gt\" (UID: \"b407ce8d-2ba4-4744-a7d3-397efcc2e0b3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5r4gt" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.143854 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-57kdl\" (UID: \"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04\") " pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.148547 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/867afc1e-15ff-4922-a2d2-5c0db0d7a019-cert\") pod \"ingress-canary-wdt5g\" (UID: \"867afc1e-15ff-4922-a2d2-5c0db0d7a019\") " pod="openshift-ingress-canary/ingress-canary-wdt5g" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.148975 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b407ce8d-2ba4-4744-a7d3-397efcc2e0b3-srv-cert\") pod \"olm-operator-6b444d44fb-5r4gt\" (UID: \"b407ce8d-2ba4-4744-a7d3-397efcc2e0b3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5r4gt" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.148962 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/717728c1-d998-4fc7-a0d2-ae90e3da4d43-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-h59nb\" (UID: \"717728c1-d998-4fc7-a0d2-ae90e3da4d43\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h59nb" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.149067 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/00adaa61-ef98-4a06-9f60-4bd990cce287-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9cjnl\" (UID: \"00adaa61-ef98-4a06-9f60-4bd990cce287\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9cjnl" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.149373 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5a3187cd-89a0-4c86-acd9-e325360446eb-proxy-tls\") pod \"machine-config-operator-74547568cd-qkhqm\" (UID: \"5a3187cd-89a0-4c86-acd9-e325360446eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qkhqm" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.149406 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a9d5fee7-2298-4320-a8d1-c1bcb17c4917-serving-cert\") pod \"service-ca-operator-777779d784-dxlk5\" (UID: \"a9d5fee7-2298-4320-a8d1-c1bcb17c4917\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dxlk5" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.152114 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/72930709-eec8-4c50-a0da-175e7cbea561-proxy-tls\") pod \"machine-config-controller-84d6567774-lb2kf\" (UID: \"72930709-eec8-4c50-a0da-175e7cbea561\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lb2kf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.152507 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swzrr\" (UniqueName: \"kubernetes.io/projected/a47192bb-5458-4644-ba17-c9873f34922d-kube-api-access-swzrr\") pod \"multus-admission-controller-857f4d67dd-9c5lc\" (UID: \"a47192bb-5458-4644-ba17-c9873f34922d\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-9c5lc" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.154444 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/9dbbbebf-e8db-40e2-9d1e-897a7b309c3a-certs\") pod \"machine-config-server-m4jm6\" (UID: \"9dbbbebf-e8db-40e2-9d1e-897a7b309c3a\") " pod="openshift-machine-config-operator/machine-config-server-m4jm6" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.168092 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gk2xq\" (UniqueName: \"kubernetes.io/projected/3f8730fb-a1a2-440e-8fb2-1bf9134054e2-kube-api-access-gk2xq\") pod \"router-default-5444994796-zd9ns\" (UID: \"3f8730fb-a1a2-440e-8fb2-1bf9134054e2\") " pod="openshift-ingress/router-default-5444994796-zd9ns" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.184225 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.200909 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l42rv\" (UniqueName: \"kubernetes.io/projected/2263dbb9-5bfb-4557-b1f2-2f33ef57db55-kube-api-access-l42rv\") pod \"kube-storage-version-migrator-operator-b67b599dd-g96sx\" (UID: \"2263dbb9-5bfb-4557-b1f2-2f33ef57db55\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g96sx" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.224543 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b819b4e5-85bd-4a31-bc2e-a119da615675-bound-sa-token\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.229377 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqbp5\" (UniqueName: \"kubernetes.io/projected/7fe6d35a-13be-4143-9072-5631cd656278-kube-api-access-kqbp5\") pod \"service-ca-9c57cc56f-6jmqc\" (UID: \"7fe6d35a-13be-4143-9072-5631cd656278\") " pod="openshift-service-ca/service-ca-9c57cc56f-6jmqc" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.237304 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:26 crc kubenswrapper[4923]: E0311 13:24:26.237754 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:26.737739052 +0000 UTC m=+114.419640766 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.251277 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k8rp\" (UniqueName: \"kubernetes.io/projected/b819b4e5-85bd-4a31-bc2e-a119da615675-kube-api-access-8k8rp\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.280632 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g96sx" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.298071 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kg2m\" (UniqueName: \"kubernetes.io/projected/714e692b-982f-4ba9-9d26-8819c83b14fe-kube-api-access-7kg2m\") pod \"dns-default-7lx27\" (UID: \"714e692b-982f-4ba9-9d26-8819c83b14fe\") " pod="openshift-dns/dns-default-7lx27" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.302173 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rr5hc\" (UniqueName: \"kubernetes.io/projected/4f20534b-c922-4867-9294-fffaddb02fbf-kube-api-access-rr5hc\") pod \"catalog-operator-68c6474976-9grj8\" (UID: \"4f20534b-c922-4867-9294-fffaddb02fbf\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9grj8" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.305798 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-l6bf9"] Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.321870 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-mgrcg"] Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.322078 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h7jfj" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.327677 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7rzzf"] Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.330621 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-7lx27" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.337695 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-zd9ns" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.338526 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:26 crc kubenswrapper[4923]: E0311 13:24:26.338855 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:26.838843592 +0000 UTC m=+114.520745296 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.342457 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c6b266fe-b225-4a79-a334-6be2b001c1b0-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6v8k4\" (UID: \"c6b266fe-b225-4a79-a334-6be2b001c1b0\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6v8k4" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.344476 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qnljg" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.350635 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9grj8" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.359489 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfpt2\" (UniqueName: \"kubernetes.io/projected/0bfac15f-c184-44be-8490-81dfe546f94f-kube-api-access-rfpt2\") pod \"ingress-operator-5b745b69d9-2c4h7\" (UID: \"0bfac15f-c184-44be-8490-81dfe546f94f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2c4h7" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.365865 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-9c5lc" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.367538 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6v8k4" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.388061 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-6jmqc" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.403732 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29553915-5tqf9" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.410567 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gppsf\" (UniqueName: \"kubernetes.io/projected/81c57b2e-1df9-44c1-9749-21fe5780f479-kube-api-access-gppsf\") pod \"packageserver-d55dfcdfc-lc5r4\" (UID: \"81c57b2e-1df9-44c1-9749-21fe5780f479\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.418682 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a7bc92e0-e42d-46f4-bc56-5123cb6d36f0-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-dqr7s\" (UID: \"a7bc92e0-e42d-46f4-bc56-5123cb6d36f0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqr7s" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.440414 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:26 crc kubenswrapper[4923]: E0311 13:24:26.440756 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:26.940741661 +0000 UTC m=+114.622643375 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.442005 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntg7g\" (UniqueName: \"kubernetes.io/projected/9dbbbebf-e8db-40e2-9d1e-897a7b309c3a-kube-api-access-ntg7g\") pod \"machine-config-server-m4jm6\" (UID: \"9dbbbebf-e8db-40e2-9d1e-897a7b309c3a\") " pod="openshift-machine-config-operator/machine-config-server-m4jm6" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.454991 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mm4gf"] Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.472506 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px8lm\" (UniqueName: \"kubernetes.io/projected/b407ce8d-2ba4-4744-a7d3-397efcc2e0b3-kube-api-access-px8lm\") pod \"olm-operator-6b444d44fb-5r4gt\" (UID: \"b407ce8d-2ba4-4744-a7d3-397efcc2e0b3\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5r4gt" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.481923 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv"] Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.495045 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh58q\" (UniqueName: \"kubernetes.io/projected/867afc1e-15ff-4922-a2d2-5c0db0d7a019-kube-api-access-nh58q\") pod \"ingress-canary-wdt5g\" (UID: \"867afc1e-15ff-4922-a2d2-5c0db0d7a019\") " pod="openshift-ingress-canary/ingress-canary-wdt5g" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.495902 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-m4jm6" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.500784 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rj5bs\" (UniqueName: \"kubernetes.io/projected/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-kube-api-access-rj5bs\") pod \"cni-sysctl-allowlist-ds-57kdl\" (UID: \"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04\") " pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.508307 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7vh5\" (UniqueName: \"kubernetes.io/projected/717728c1-d998-4fc7-a0d2-ae90e3da4d43-kube-api-access-s7vh5\") pod \"package-server-manager-789f6589d5-h59nb\" (UID: \"717728c1-d998-4fc7-a0d2-ae90e3da4d43\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h59nb" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.531641 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rl5p\" (UniqueName: \"kubernetes.io/projected/b0a9d170-2b10-48f1-81ab-898a7df26740-kube-api-access-8rl5p\") pod \"csi-hostpathplugin-lc9xf\" (UID: \"b0a9d170-2b10-48f1-81ab-898a7df26740\") " pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.543786 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:26 crc kubenswrapper[4923]: E0311 13:24:26.544165 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:27.044152828 +0000 UTC m=+114.726054542 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.556637 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-wdt5g" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.561900 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8jkj\" (UniqueName: \"kubernetes.io/projected/5a3187cd-89a0-4c86-acd9-e325360446eb-kube-api-access-w8jkj\") pod \"machine-config-operator-74547568cd-qkhqm\" (UID: \"5a3187cd-89a0-4c86-acd9-e325360446eb\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qkhqm" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.573277 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6x696\" (UniqueName: \"kubernetes.io/projected/a9d5fee7-2298-4320-a8d1-c1bcb17c4917-kube-api-access-6x696\") pod \"service-ca-operator-777779d784-dxlk5\" (UID: \"a9d5fee7-2298-4320-a8d1-c1bcb17c4917\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dxlk5" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.613950 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j77fn\" (UniqueName: \"kubernetes.io/projected/00adaa61-ef98-4a06-9f60-4bd990cce287-kube-api-access-j77fn\") pod \"control-plane-machine-set-operator-78cbb6b69f-9cjnl\" (UID: \"00adaa61-ef98-4a06-9f60-4bd990cce287\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9cjnl" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.614528 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58rk8\" (UniqueName: \"kubernetes.io/projected/72930709-eec8-4c50-a0da-175e7cbea561-kube-api-access-58rk8\") pod \"machine-config-controller-84d6567774-lb2kf\" (UID: \"72930709-eec8-4c50-a0da-175e7cbea561\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lb2kf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.616943 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gfx4\" (UniqueName: \"kubernetes.io/projected/b6887dd3-ec66-4364-8eed-f63615c35024-kube-api-access-9gfx4\") pod \"marketplace-operator-79b997595-vznqg\" (UID: \"b6887dd3-ec66-4364-8eed-f63615c35024\") " pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.623068 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2c4h7" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.641968 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8bhpz"] Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.645311 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:26 crc kubenswrapper[4923]: E0311 13:24:26.659508 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:27.145390541 +0000 UTC m=+114.827292255 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.663207 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:26 crc kubenswrapper[4923]: E0311 13:24:26.663681 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:27.163662275 +0000 UTC m=+114.845563989 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.677244 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqr7s" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.691721 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.709967 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5r4gt" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.733095 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9cjnl" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.742723 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dxlk5" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.742900 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h59nb" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.742769 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.742841 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lb2kf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.748308 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qkhqm" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.757096 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.766832 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:26 crc kubenswrapper[4923]: E0311 13:24:26.768807 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:27.268785874 +0000 UTC m=+114.950687588 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.769417 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:26 crc kubenswrapper[4923]: E0311 13:24:26.769806 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:27.26979816 +0000 UTC m=+114.951699874 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.773445 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" event={"ID":"a67361c9-12cb-4a1a-893c-a735a58da9cd","Type":"ContainerStarted","Data":"fb0494a4e891c50c95ba68bfe20613568c1d2b4fe6a21266b4098b86f4beb86f"} Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.773931 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hgkc2"] Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.784421 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-k6dxb"] Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.797108 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.812074 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l6bf9" event={"ID":"919663b1-49e2-4405-b62a-5cd8b7092953","Type":"ContainerStarted","Data":"9edbabfb3050fc73dfb6016c47d559752328aae23195a2276115e0c8ebaffaa4"} Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.815114 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jfhp5" event={"ID":"5ec8d4f8-79c7-4153-94f1-0efe16a3da0f","Type":"ContainerStarted","Data":"37ef64449b7fa25c8e13e97a39031355ee3c78483ddb37b035e53379b6098eea"} Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.841122 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-sdtpq" event={"ID":"a6e6cf79-4d28-4262-801f-f85de17c5def","Type":"ContainerStarted","Data":"af92c056d27b36840193162cdbc1f111753e3e77b00a007ebe1de730a06dc197"} Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.843748 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-zd9ns" event={"ID":"3f8730fb-a1a2-440e-8fb2-1bf9134054e2","Type":"ContainerStarted","Data":"49569fcae916f4edd83dffc321a11e6aac633c268efeedf0190521efc1d1ada2"} Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.865669 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" event={"ID":"7c496315-8928-482f-be56-265af63f3888","Type":"ContainerStarted","Data":"040ca14c065b6dd5bb6d151ccf2b0887cf24c126c2afc1e3cd83590c29aefb8e"} Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.876694 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:26 crc kubenswrapper[4923]: E0311 13:24:26.876972 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:27.37695638 +0000 UTC m=+115.058858094 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.880700 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pk9tb"] Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.883450 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mm4gf" event={"ID":"15bb3d2b-b085-4cf9-8ea1-1a67a5bc74e1","Type":"ContainerStarted","Data":"7ea79566a979cd7299c86fc595e66c996d23ef068e5a48a3e0556b4aae2ec143"} Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.916579 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" event={"ID":"cfda8f36-49b9-4081-bbea-f311dd0a87a1","Type":"ContainerStarted","Data":"a23af8b5446190635a930538e80a26c355ca590fd3665b3e9da27144b5ad9381"} Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.916620 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" event={"ID":"cfda8f36-49b9-4081-bbea-f311dd0a87a1","Type":"ContainerStarted","Data":"5e3664da06955918680e8c88d95740647f4ed9785e064e41f9e6b45a84b01679"} Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.917528 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.925667 4923 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-qk4ql container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.925738 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" podUID="cfda8f36-49b9-4081-bbea-f311dd0a87a1" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.931735 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mgrcg" event={"ID":"13238a87-3b9a-471a-8337-7e7f15018221","Type":"ContainerStarted","Data":"15d08441769e153af24275a28743b328ac9abd6132f53813aa1cd4c14b28361d"} Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.937425 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g96sx"] Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.972226 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-4nhkq" podStartSLOduration=61.972209444 podStartE2EDuration="1m1.972209444s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:26.970221075 +0000 UTC m=+114.652122789" watchObservedRunningTime="2026-03-11 13:24:26.972209444 +0000 UTC m=+114.654111158" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.978631 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:26 crc kubenswrapper[4923]: E0311 13:24:26.979119 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:27.479107695 +0000 UTC m=+115.161009409 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.980178 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"f999e7fa000b30bfc260bf58933882566bfe11273e586fcbf28dc0a699a176e5"} Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.988456 4923 patch_prober.go:28] interesting pod/downloads-7954f5f757-b5jmc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" start-of-body= Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.988504 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b5jmc" podUID="c7215b86-4f1e-44fb-b0a9-c7d261ce4a54" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" Mar 11 13:24:26 crc kubenswrapper[4923]: I0311 13:24:26.992734 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" Mar 11 13:24:27 crc kubenswrapper[4923]: I0311 13:24:27.081387 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:27 crc kubenswrapper[4923]: E0311 13:24:27.082880 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:27.582864111 +0000 UTC m=+115.264765825 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:27 crc kubenswrapper[4923]: I0311 13:24:27.185008 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:27 crc kubenswrapper[4923]: E0311 13:24:27.185772 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:27.685751135 +0000 UTC m=+115.367652849 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:27 crc kubenswrapper[4923]: W0311 13:24:27.192676 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4890044b_8d18_4123_b029_7e2809294be1.slice/crio-c6ad46aee2639f054c856ab05246cecdfafbe479d7c09d18f4feac16c54bf3ff WatchSource:0}: Error finding container c6ad46aee2639f054c856ab05246cecdfafbe479d7c09d18f4feac16c54bf3ff: Status 404 returned error can't find the container with id c6ad46aee2639f054c856ab05246cecdfafbe479d7c09d18f4feac16c54bf3ff Mar 11 13:24:27 crc kubenswrapper[4923]: I0311 13:24:27.299808 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:27 crc kubenswrapper[4923]: E0311 13:24:27.300422 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:27.800401261 +0000 UTC m=+115.482302975 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:27 crc kubenswrapper[4923]: W0311 13:24:27.350238 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-cb148a2750b711fa0cd43093e003380bb21286651ae892c31f1b343d045fc5af WatchSource:0}: Error finding container cb148a2750b711fa0cd43093e003380bb21286651ae892c31f1b343d045fc5af: Status 404 returned error can't find the container with id cb148a2750b711fa0cd43093e003380bb21286651ae892c31f1b343d045fc5af Mar 11 13:24:27 crc kubenswrapper[4923]: I0311 13:24:27.372980 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-7lx27"] Mar 11 13:24:27 crc kubenswrapper[4923]: I0311 13:24:27.403678 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:27 crc kubenswrapper[4923]: E0311 13:24:27.403981 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:27.903968602 +0000 UTC m=+115.585870316 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:27 crc kubenswrapper[4923]: I0311 13:24:27.448506 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-c7dks" podStartSLOduration=62.448492307 podStartE2EDuration="1m2.448492307s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:27.4482009 +0000 UTC m=+115.130102624" watchObservedRunningTime="2026-03-11 13:24:27.448492307 +0000 UTC m=+115.130394021" Mar 11 13:24:27 crc kubenswrapper[4923]: I0311 13:24:27.506992 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9grj8"] Mar 11 13:24:27 crc kubenswrapper[4923]: I0311 13:24:27.507462 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:27 crc kubenswrapper[4923]: E0311 13:24:27.507735 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:28.007721697 +0000 UTC m=+115.689623411 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:27 crc kubenswrapper[4923]: I0311 13:24:27.534077 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-9c5lc"] Mar 11 13:24:27 crc kubenswrapper[4923]: I0311 13:24:27.535906 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-h7jfj"] Mar 11 13:24:27 crc kubenswrapper[4923]: I0311 13:24:27.608941 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:27 crc kubenswrapper[4923]: E0311 13:24:27.609287 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:28.109270078 +0000 UTC m=+115.791171792 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:27 crc kubenswrapper[4923]: I0311 13:24:27.708692 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" podStartSLOduration=62.708673076 podStartE2EDuration="1m2.708673076s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:27.655045215 +0000 UTC m=+115.336946939" watchObservedRunningTime="2026-03-11 13:24:27.708673076 +0000 UTC m=+115.390574790" Mar 11 13:24:27 crc kubenswrapper[4923]: I0311 13:24:27.710908 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:27 crc kubenswrapper[4923]: E0311 13:24:27.711254 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:28.21123908 +0000 UTC m=+115.893140794 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:27 crc kubenswrapper[4923]: I0311 13:24:27.735601 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-b5jmc" podStartSLOduration=62.735584293 podStartE2EDuration="1m2.735584293s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:27.734485907 +0000 UTC m=+115.416387621" watchObservedRunningTime="2026-03-11 13:24:27.735584293 +0000 UTC m=+115.417486007" Mar 11 13:24:27 crc kubenswrapper[4923]: W0311 13:24:27.749679 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e0c6535_0c45_4d8d_9b97_332888787b98.slice/crio-44265e6eebd427878e2c3833e4877301bada8bbd6d77cd72cebf8a73871caaf8 WatchSource:0}: Error finding container 44265e6eebd427878e2c3833e4877301bada8bbd6d77cd72cebf8a73871caaf8: Status 404 returned error can't find the container with id 44265e6eebd427878e2c3833e4877301bada8bbd6d77cd72cebf8a73871caaf8 Mar 11 13:24:27 crc kubenswrapper[4923]: I0311 13:24:27.812041 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:27 crc kubenswrapper[4923]: E0311 13:24:27.812386 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:28.31237503 +0000 UTC m=+115.994276744 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:27 crc kubenswrapper[4923]: I0311 13:24:27.908805 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-r2t4r" podStartSLOduration=62.908783513 podStartE2EDuration="1m2.908783513s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:27.871461147 +0000 UTC m=+115.553362861" watchObservedRunningTime="2026-03-11 13:24:27.908783513 +0000 UTC m=+115.590685227" Mar 11 13:24:27 crc kubenswrapper[4923]: I0311 13:24:27.910021 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-6jmqc"] Mar 11 13:24:27 crc kubenswrapper[4923]: I0311 13:24:27.916925 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:27 crc kubenswrapper[4923]: E0311 13:24:27.917193 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:28.417178061 +0000 UTC m=+116.099079775 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:27 crc kubenswrapper[4923]: I0311 13:24:27.930415 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29553915-5tqf9"] Mar 11 13:24:27 crc kubenswrapper[4923]: I0311 13:24:27.935792 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" podStartSLOduration=62.935765253 podStartE2EDuration="1m2.935765253s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:27.921275554 +0000 UTC m=+115.603177268" watchObservedRunningTime="2026-03-11 13:24:27.935765253 +0000 UTC m=+115.617666967" Mar 11 13:24:27 crc kubenswrapper[4923]: I0311 13:24:27.941029 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qnljg"] Mar 11 13:24:27 crc kubenswrapper[4923]: I0311 13:24:27.941069 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6v8k4"] Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.021877 4923 generic.go:334] "Generic (PLEG): container finished" podID="919663b1-49e2-4405-b62a-5cd8b7092953" containerID="70fd9d0efa83087f73316433828a6fc716dd41747beb625290ac8d15db7577b2" exitCode=0 Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.022215 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l6bf9" event={"ID":"919663b1-49e2-4405-b62a-5cd8b7092953","Type":"ContainerDied","Data":"70fd9d0efa83087f73316433828a6fc716dd41747beb625290ac8d15db7577b2"} Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.023028 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:28 crc kubenswrapper[4923]: E0311 13:24:28.023286 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:28.523275215 +0000 UTC m=+116.205176929 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.034945 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pk9tb" event={"ID":"4890044b-8d18-4123-b029-7e2809294be1","Type":"ContainerStarted","Data":"c6ad46aee2639f054c856ab05246cecdfafbe479d7c09d18f4feac16c54bf3ff"} Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.081474 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" event={"ID":"1515573d-f360-41f2-854a-4d0fb437af7d","Type":"ContainerStarted","Data":"8d20467cdf5be8d05fd96e6e843500c18ef9b395d23c5e96a79ee5168b705457"} Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.096983 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-ndw7p" podStartSLOduration=63.096964634 podStartE2EDuration="1m3.096964634s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:28.066614531 +0000 UTC m=+115.748516245" watchObservedRunningTime="2026-03-11 13:24:28.096964634 +0000 UTC m=+115.778866348" Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.108733 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-zd9ns" event={"ID":"3f8730fb-a1a2-440e-8fb2-1bf9134054e2","Type":"ContainerStarted","Data":"6a52107d61651cac27b7a150017eb8f69ee05ed25fbebdbc7d5cf18f5065d125"} Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.123962 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:28 crc kubenswrapper[4923]: E0311 13:24:28.125143 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:28.625120453 +0000 UTC m=+116.307022167 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.131517 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-7lx27" event={"ID":"714e692b-982f-4ba9-9d26-8819c83b14fe","Type":"ContainerStarted","Data":"21397c44c98aff10b3630ea2db698efb660be7ba6171a6595b9f725b79b9b47e"} Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.150751 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-9c5lc" event={"ID":"a47192bb-5458-4644-ba17-c9873f34922d","Type":"ContainerStarted","Data":"5dceab14488c778906efa02ad776650f62fe63a64b201096406eb769cd31585b"} Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.175788 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-wdt5g"] Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.226418 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:28 crc kubenswrapper[4923]: E0311 13:24:28.227840 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:28.727823423 +0000 UTC m=+116.409725137 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.228539 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-m4jm6" event={"ID":"9dbbbebf-e8db-40e2-9d1e-897a7b309c3a","Type":"ContainerStarted","Data":"1a96065a57be16ba7bed94a9d6f0be3e68d700990be81fe31026d1b1c410a2ee"} Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.228571 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-m4jm6" event={"ID":"9dbbbebf-e8db-40e2-9d1e-897a7b309c3a","Type":"ContainerStarted","Data":"9025f5b8164e02cd51d8bd3d331ded618777d3f36154afb2322b6c551ca93b7b"} Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.252585 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4"] Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.288418 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-lc9xf"] Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.327559 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vznqg"] Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.327599 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8bhpz" event={"ID":"f4826a3f-d90e-4b7c-a10b-864b78115463","Type":"ContainerStarted","Data":"dedb1d58ad966a21a52b128df8b7347e002bec66c3bf202917bac2ebc4563163"} Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.328191 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:28 crc kubenswrapper[4923]: E0311 13:24:28.328386 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:28.828371779 +0000 UTC m=+116.510273493 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.328471 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:28 crc kubenswrapper[4923]: E0311 13:24:28.329324 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:28.829310882 +0000 UTC m=+116.511212596 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.340543 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-zd9ns" Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.344949 4923 patch_prober.go:28] interesting pod/router-default-5444994796-zd9ns container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.344996 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zd9ns" podUID="3f8730fb-a1a2-440e-8fb2-1bf9134054e2" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.347385 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g96sx" event={"ID":"2263dbb9-5bfb-4557-b1f2-2f33ef57db55","Type":"ContainerStarted","Data":"630c7e5ef45508e65928ef1c012642eeffe950ed7ac889213e432f8f05770fd8"} Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.372325 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9grj8" event={"ID":"4f20534b-c922-4867-9294-fffaddb02fbf","Type":"ContainerStarted","Data":"f3329b7259d0c8e3014f1a96213b5fd764a6931861b6ef03b6d29c2c92c7cac0"} Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.382815 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h59nb"] Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.384117 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-qkhqm"] Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.412685 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-2c4h7"] Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.413094 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqr7s"] Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.415211 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-m4jm6" podStartSLOduration=5.415195604 podStartE2EDuration="5.415195604s" podCreationTimestamp="2026-03-11 13:24:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:28.4021259 +0000 UTC m=+116.084027614" watchObservedRunningTime="2026-03-11 13:24:28.415195604 +0000 UTC m=+116.097097318" Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.429263 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:28 crc kubenswrapper[4923]: E0311 13:24:28.429561 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:28.929546201 +0000 UTC m=+116.611447905 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.449761 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-sdtpq" event={"ID":"a6e6cf79-4d28-4262-801f-f85de17c5def","Type":"ContainerStarted","Data":"0c0fe286951d9c64c9df8b3243b6428f61b449b23dbc0d71cabff8c11a1b8c85"} Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.451671 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-sdtpq" Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.455571 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-lb2kf"] Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.456197 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h7jfj" event={"ID":"7e0c6535-0c45-4d8d-9b97-332888787b98","Type":"ContainerStarted","Data":"44265e6eebd427878e2c3833e4877301bada8bbd6d77cd72cebf8a73871caaf8"} Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.457361 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-zd9ns" podStartSLOduration=63.4573225 podStartE2EDuration="1m3.4573225s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:28.44563999 +0000 UTC m=+116.127541704" watchObservedRunningTime="2026-03-11 13:24:28.4573225 +0000 UTC m=+116.139224214" Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.462952 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-dxlk5"] Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.474392 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"cb148a2750b711fa0cd43093e003380bb21286651ae892c31f1b343d045fc5af"} Mar 11 13:24:28 crc kubenswrapper[4923]: W0311 13:24:28.492277 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod717728c1_d998_4fc7_a0d2_ae90e3da4d43.slice/crio-a566c1e8922b5441ecd1e9595e2949a728e32708fdb4aa14ddd178428b7caa24 WatchSource:0}: Error finding container a566c1e8922b5441ecd1e9595e2949a728e32708fdb4aa14ddd178428b7caa24: Status 404 returned error can't find the container with id a566c1e8922b5441ecd1e9595e2949a728e32708fdb4aa14ddd178428b7caa24 Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.511663 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" event={"ID":"7c496315-8928-482f-be56-265af63f3888","Type":"ContainerStarted","Data":"ce57ac1585e2c81bc23be72eebf9aa74b6b0ee4c22a35dd0dd1aeecbcd1e4cce"} Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.513767 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5r4gt"] Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.519962 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9cjnl"] Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.535262 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:28 crc kubenswrapper[4923]: E0311 13:24:28.537515 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:29.037497831 +0000 UTC m=+116.719399545 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.540092 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" event={"ID":"08068104-34d8-440d-8f98-f9a6216af858","Type":"ContainerStarted","Data":"41074d26a93e59829551b8f1b148bd83529020e9b7f09d8e91e43fcb0c06c711"} Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.547313 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jfhp5" event={"ID":"5ec8d4f8-79c7-4153-94f1-0efe16a3da0f","Type":"ContainerStarted","Data":"2d5269b8c8a18fd01916dff30c940553255b7219ee01435cf57e11544dbf1578"} Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.559675 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" event={"ID":"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04","Type":"ContainerStarted","Data":"1de5ae939254de4f0275f4db1cc697cb33c414e6163743ef45a0461bf150cd47"} Mar 11 13:24:28 crc kubenswrapper[4923]: W0311 13:24:28.560604 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb407ce8d_2ba4_4744_a7d3_397efcc2e0b3.slice/crio-693e1763e988e50097ab74fc20fc88195858db36b1d2ee42aff5a6121479fdf3 WatchSource:0}: Error finding container 693e1763e988e50097ab74fc20fc88195858db36b1d2ee42aff5a6121479fdf3: Status 404 returned error can't find the container with id 693e1763e988e50097ab74fc20fc88195858db36b1d2ee42aff5a6121479fdf3 Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.560668 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.562790 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-sdtpq" podStartSLOduration=63.562769688 podStartE2EDuration="1m3.562769688s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:28.533077351 +0000 UTC m=+116.214979065" watchObservedRunningTime="2026-03-11 13:24:28.562769688 +0000 UTC m=+116.244671402" Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.576612 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"f1f9aed14c2bef16a3d3834a091c8e29c7bb531a0ee246aeed17ac0da223847d"} Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.583791 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.635843 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:28 crc kubenswrapper[4923]: E0311 13:24:28.640585 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:29.140562488 +0000 UTC m=+116.822464202 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.659637 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jfhp5" podStartSLOduration=63.659621562 podStartE2EDuration="1m3.659621562s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:28.608128884 +0000 UTC m=+116.290030598" watchObservedRunningTime="2026-03-11 13:24:28.659621562 +0000 UTC m=+116.341523276" Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.682051 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" podStartSLOduration=5.682031908 podStartE2EDuration="5.682031908s" podCreationTimestamp="2026-03-11 13:24:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:28.681791012 +0000 UTC m=+116.363692726" watchObservedRunningTime="2026-03-11 13:24:28.682031908 +0000 UTC m=+116.363933632" Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.685363 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-7rzzf" podStartSLOduration=63.68533664 podStartE2EDuration="1m3.68533664s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:28.660331539 +0000 UTC m=+116.342233253" watchObservedRunningTime="2026-03-11 13:24:28.68533664 +0000 UTC m=+116.367238374" Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.737465 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:28 crc kubenswrapper[4923]: E0311 13:24:28.738570 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:29.238554651 +0000 UTC m=+116.920456365 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.741316 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-sdtpq" Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.847592 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:28 crc kubenswrapper[4923]: E0311 13:24:28.847988 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:29.347956327 +0000 UTC m=+117.029858041 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.848407 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:28 crc kubenswrapper[4923]: E0311 13:24:28.848739 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:29.348722976 +0000 UTC m=+117.030624690 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.949149 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.949673 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ae7e45f7-8c08-435b-990a-5149da33010a-metrics-certs\") pod \"network-metrics-daemon-4nx54\" (UID: \"ae7e45f7-8c08-435b-990a-5149da33010a\") " pod="openshift-multus/network-metrics-daemon-4nx54" Mar 11 13:24:28 crc kubenswrapper[4923]: E0311 13:24:28.949882 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:29.449850857 +0000 UTC m=+117.131752571 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:28 crc kubenswrapper[4923]: I0311 13:24:28.973266 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ae7e45f7-8c08-435b-990a-5149da33010a-metrics-certs\") pod \"network-metrics-daemon-4nx54\" (UID: \"ae7e45f7-8c08-435b-990a-5149da33010a\") " pod="openshift-multus/network-metrics-daemon-4nx54" Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.050547 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:29 crc kubenswrapper[4923]: E0311 13:24:29.051080 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:29.551055999 +0000 UTC m=+117.232957723 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.151579 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:29 crc kubenswrapper[4923]: E0311 13:24:29.151913 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:29.651897602 +0000 UTC m=+117.333799316 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.236535 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4nx54" Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.253579 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:29 crc kubenswrapper[4923]: E0311 13:24:29.253929 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:29.753917034 +0000 UTC m=+117.435818748 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.347393 4923 patch_prober.go:28] interesting pod/router-default-5444994796-zd9ns container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 11 13:24:29 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Mar 11 13:24:29 crc kubenswrapper[4923]: [+]process-running ok Mar 11 13:24:29 crc kubenswrapper[4923]: healthz check failed Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.347441 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zd9ns" podUID="3f8730fb-a1a2-440e-8fb2-1bf9134054e2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.355138 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:29 crc kubenswrapper[4923]: E0311 13:24:29.355572 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:29.855541318 +0000 UTC m=+117.537443032 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.457215 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:29 crc kubenswrapper[4923]: E0311 13:24:29.458139 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:29.958106063 +0000 UTC m=+117.640007777 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.558923 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:29 crc kubenswrapper[4923]: E0311 13:24:29.559739 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:30.059718105 +0000 UTC m=+117.741619819 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.610369 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pk9tb" event={"ID":"4890044b-8d18-4123-b029-7e2809294be1","Type":"ContainerStarted","Data":"338e18311cb9257203a14d6488efadb75d3b55a520de31e2bf1957c0cafdb91f"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.635738 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-4nx54"] Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.658001 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-jfhp5" event={"ID":"5ec8d4f8-79c7-4153-94f1-0efe16a3da0f","Type":"ContainerStarted","Data":"0a7cf150146950a997a564b1c602001d242b0c2ff29aa0c9bc1775fb658ec7e3"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.660536 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:29 crc kubenswrapper[4923]: E0311 13:24:29.660936 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:30.160916828 +0000 UTC m=+117.842818552 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.679410 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-pk9tb" podStartSLOduration=64.679392416 podStartE2EDuration="1m4.679392416s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:29.677265173 +0000 UTC m=+117.359166887" watchObservedRunningTime="2026-03-11 13:24:29.679392416 +0000 UTC m=+117.361294130" Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.684365 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lb2kf" event={"ID":"72930709-eec8-4c50-a0da-175e7cbea561","Type":"ContainerStarted","Data":"7f4f069a46139e8e4b2ca1c7d846f585088040c36b5d982d69e89b29dce8ed99"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.684477 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lb2kf" event={"ID":"72930709-eec8-4c50-a0da-175e7cbea561","Type":"ContainerStarted","Data":"c6d1ce52a1f03709a017d1fe8460a8e5195390974e3191621fc4305633c7517d"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.735422 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqr7s" event={"ID":"a7bc92e0-e42d-46f4-bc56-5123cb6d36f0","Type":"ContainerStarted","Data":"e9bfefa6f30b9534bf6e9c2f393706f30b9954f15e0620c6aabb6f830e7af851"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.761835 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:29 crc kubenswrapper[4923]: E0311 13:24:29.763147 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:30.263127235 +0000 UTC m=+117.945028949 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.777007 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5r4gt" event={"ID":"b407ce8d-2ba4-4744-a7d3-397efcc2e0b3","Type":"ContainerStarted","Data":"1ed65dc213e9e4ba3630c7f5559f21b83ee85b00dfaafad5d9345bcf0e36bc8a"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.777060 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5r4gt" event={"ID":"b407ce8d-2ba4-4744-a7d3-397efcc2e0b3","Type":"ContainerStarted","Data":"693e1763e988e50097ab74fc20fc88195858db36b1d2ee42aff5a6121479fdf3"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.778084 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5r4gt" Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.791137 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29553915-5tqf9" event={"ID":"122f383b-c9fa-4ab3-8347-e75d16a0189b","Type":"ContainerStarted","Data":"518ed8856b0f90fde2151e77d68426004300ca1a92498bd44fae594d2cecc338"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.791208 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29553915-5tqf9" event={"ID":"122f383b-c9fa-4ab3-8347-e75d16a0189b","Type":"ContainerStarted","Data":"9cc6f13e7b62558bf379c1647be8887c5fa2a19760e4486387981eb7d4784395"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.795328 4923 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-5r4gt container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" start-of-body= Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.795449 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5r4gt" podUID="b407ce8d-2ba4-4744-a7d3-397efcc2e0b3" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.816449 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2c4h7" event={"ID":"0bfac15f-c184-44be-8490-81dfe546f94f","Type":"ContainerStarted","Data":"05293f4ef0f2b70aa1e48c33bb0eb96d9e8233f432403ab9b4b5086cdea689bb"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.816500 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2c4h7" event={"ID":"0bfac15f-c184-44be-8490-81dfe546f94f","Type":"ContainerStarted","Data":"2f55489c615e9583469aa480c735c2ecd4ec320272960a67e813a1681dd063ef"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.823069 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqr7s" podStartSLOduration=64.823050932 podStartE2EDuration="1m4.823050932s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:29.791235493 +0000 UTC m=+117.473137207" watchObservedRunningTime="2026-03-11 13:24:29.823050932 +0000 UTC m=+117.504952646" Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.825198 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5r4gt" podStartSLOduration=64.825185635 podStartE2EDuration="1m4.825185635s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:29.822754945 +0000 UTC m=+117.504656669" watchObservedRunningTime="2026-03-11 13:24:29.825185635 +0000 UTC m=+117.507087349" Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.836288 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qnljg" event={"ID":"4ecd0070-8225-4da9-a577-4382d935f74b","Type":"ContainerStarted","Data":"c055f30802db6c92f2475ab50d4c90644874feb58f0c0b0288ed677fe7880add"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.836358 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qnljg" event={"ID":"4ecd0070-8225-4da9-a577-4382d935f74b","Type":"ContainerStarted","Data":"e1d0ca1e4972a6cee63440c600c64aac492ded78fab6e3bcef46e4d9d0e705b0"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.841888 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9cjnl" event={"ID":"00adaa61-ef98-4a06-9f60-4bd990cce287","Type":"ContainerStarted","Data":"2743534329eda6aa1179656e08a702f2236580c9160727268b0d3426579d5785"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.841936 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9cjnl" event={"ID":"00adaa61-ef98-4a06-9f60-4bd990cce287","Type":"ContainerStarted","Data":"69a6fb0040df4c81cc77a48a59c4286fd895f2c1f01bd4f184c98f350105bf8b"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.846776 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-9c5lc" event={"ID":"a47192bb-5458-4644-ba17-c9873f34922d","Type":"ContainerStarted","Data":"9e5b9f1b9836f2cadc64b5d70c25ca1e26953a1582e66ef007a0069cda5cab14"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.846825 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-9c5lc" event={"ID":"a47192bb-5458-4644-ba17-c9873f34922d","Type":"ContainerStarted","Data":"b9cccf65c8c93d6cde49604435a33e7038d02923f7827b26b582a5cb58aa3151"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.854702 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mm4gf" event={"ID":"15bb3d2b-b085-4cf9-8ea1-1a67a5bc74e1","Type":"ContainerStarted","Data":"34166fdfd6a3e47616a4ce8b3c5d8bffcd39f0daa6275204b4c34cea469fc786"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.855744 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29553915-5tqf9" podStartSLOduration=65.855730874 podStartE2EDuration="1m5.855730874s" podCreationTimestamp="2026-03-11 13:23:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:29.853662322 +0000 UTC m=+117.535564036" watchObservedRunningTime="2026-03-11 13:24:29.855730874 +0000 UTC m=+117.537632588" Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.865568 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:29 crc kubenswrapper[4923]: E0311 13:24:29.866141 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:30.366128022 +0000 UTC m=+118.048029736 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.876078 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4" event={"ID":"81c57b2e-1df9-44c1-9749-21fe5780f479","Type":"ContainerStarted","Data":"f5ad8adce74461bfeaee5698246aac88636174ee3f6a19b77961097fef9235e7"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.876120 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4" event={"ID":"81c57b2e-1df9-44c1-9749-21fe5780f479","Type":"ContainerStarted","Data":"470704a94d2e56b0e212193e7952e862a530e29a42dc4b7e6966ecf7df5a0ee5"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.877083 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4" Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.877350 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2c4h7" podStartSLOduration=64.87731146 podStartE2EDuration="1m4.87731146s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:29.874695995 +0000 UTC m=+117.556597709" watchObservedRunningTime="2026-03-11 13:24:29.87731146 +0000 UTC m=+117.559213174" Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.889488 4923 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-lc5r4 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:5443/healthz\": dial tcp 10.217.0.31:5443: connect: connection refused" start-of-body= Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.889549 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4" podUID="81c57b2e-1df9-44c1-9749-21fe5780f479" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.31:5443/healthz\": dial tcp 10.217.0.31:5443: connect: connection refused" Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.890796 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e84ac7634e19068c9196d35a317083bcecead2ae2cfc93a44dee794635874d53"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.896099 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qnljg" podStartSLOduration=64.896084866 podStartE2EDuration="1m4.896084866s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:29.895672585 +0000 UTC m=+117.577574299" watchObservedRunningTime="2026-03-11 13:24:29.896084866 +0000 UTC m=+117.577986580" Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.900210 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h59nb" event={"ID":"717728c1-d998-4fc7-a0d2-ae90e3da4d43","Type":"ContainerStarted","Data":"4c1b6f0df727a94010ababe4679a20746459843f8a29aa5b6d0a1deeb1aa9cb8"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.900254 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h59nb" event={"ID":"717728c1-d998-4fc7-a0d2-ae90e3da4d43","Type":"ContainerStarted","Data":"a566c1e8922b5441ecd1e9595e2949a728e32708fdb4aa14ddd178428b7caa24"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.900866 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h59nb" Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.906655 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6v8k4" event={"ID":"c6b266fe-b225-4a79-a334-6be2b001c1b0","Type":"ContainerStarted","Data":"b9d79f4fa34e0d31da488f3960953d687e092aefd2d22bb0c68d3196ad903267"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.906698 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6v8k4" event={"ID":"c6b266fe-b225-4a79-a334-6be2b001c1b0","Type":"ContainerStarted","Data":"c12fd8a9ff6a3c48a33c307107139949186561fab361287711d81e92827538c8"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.915352 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" event={"ID":"b0a9d170-2b10-48f1-81ab-898a7df26740","Type":"ContainerStarted","Data":"fcf397ac3c4ae9d7b37a9a660512cbdb8cda918d1aa5f181afb95fd55e87b1b2"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.916410 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-wdt5g" event={"ID":"867afc1e-15ff-4922-a2d2-5c0db0d7a019","Type":"ContainerStarted","Data":"649daf0f5037dfb67a96a3142f09555b5551310d49ff56bfdde4069f88c5f2a2"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.916438 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-wdt5g" event={"ID":"867afc1e-15ff-4922-a2d2-5c0db0d7a019","Type":"ContainerStarted","Data":"e364a454e97e066f0d80848c522a9ef332cd139e9d62508b36bbd8a4e5099a2a"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.923088 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9cjnl" podStartSLOduration=64.923071275 podStartE2EDuration="1m4.923071275s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:29.917161499 +0000 UTC m=+117.599063213" watchObservedRunningTime="2026-03-11 13:24:29.923071275 +0000 UTC m=+117.604972989" Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.942088 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l6bf9" event={"ID":"919663b1-49e2-4405-b62a-5cd8b7092953","Type":"ContainerStarted","Data":"0fc85b2755476c2206878bf7f2faab1ffac3832ac6188e90fb21df1136aa5575"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.942667 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l6bf9" Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.956237 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"cbf3ffe8289cf0f95b47725b0122bf2613222bad09636a7a9d032c75dc946658"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.960582 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-9c5lc" podStartSLOduration=64.960565577 podStartE2EDuration="1m4.960565577s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:29.939052422 +0000 UTC m=+117.620954146" watchObservedRunningTime="2026-03-11 13:24:29.960565577 +0000 UTC m=+117.642467291" Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.961433 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6v8k4" podStartSLOduration=64.961427688 podStartE2EDuration="1m4.961427688s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:29.959746246 +0000 UTC m=+117.641647960" watchObservedRunningTime="2026-03-11 13:24:29.961427688 +0000 UTC m=+117.643329392" Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.961605 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-6jmqc" event={"ID":"7fe6d35a-13be-4143-9072-5631cd656278","Type":"ContainerStarted","Data":"e98c6e184f70efed73198cc5941d2169ba0a0c77167497cc951b16c853e1e3e5"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.961651 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-6jmqc" event={"ID":"7fe6d35a-13be-4143-9072-5631cd656278","Type":"ContainerStarted","Data":"043ce483647295d50fc37be5411b74bbf290302e21332b46c22bfa389f383c30"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.966768 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:29 crc kubenswrapper[4923]: E0311 13:24:29.967910 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:30.467890208 +0000 UTC m=+118.149791922 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.976504 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" event={"ID":"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04","Type":"ContainerStarted","Data":"d348f00e8a6e4b07e7a23195b2d70fdf885db23f4ce1b740a3fcb77ccd7756ec"} Mar 11 13:24:29 crc kubenswrapper[4923]: I0311 13:24:29.989257 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-wdt5g" podStartSLOduration=6.989236308 podStartE2EDuration="6.989236308s" podCreationTimestamp="2026-03-11 13:24:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:29.985273219 +0000 UTC m=+117.667174933" watchObservedRunningTime="2026-03-11 13:24:29.989236308 +0000 UTC m=+117.671138042" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:29.999940 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qkhqm" event={"ID":"5a3187cd-89a0-4c86-acd9-e325360446eb","Type":"ContainerStarted","Data":"590e1d9ab4cf4019e5316f2246c45da089a94b2e93f404fca522b7bd4303f112"} Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:29.999990 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qkhqm" event={"ID":"5a3187cd-89a0-4c86-acd9-e325360446eb","Type":"ContainerStarted","Data":"0e3fa9c00baa6830241f1303239de8d48f9a1a9beda52ae76a5695ed63394bc2"} Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.003827 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mgrcg" event={"ID":"13238a87-3b9a-471a-8337-7e7f15018221","Type":"ContainerStarted","Data":"e46a649a6a0f894fd07fe2be7416d92d0031604243845aa1241a916eb68fb520"} Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.010006 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"5086acc8a0412e708e722aa7ae898e664e5240289a56931de7b307f5d6156948"} Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.010860 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.022884 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h59nb" podStartSLOduration=65.022854783 podStartE2EDuration="1m5.022854783s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:30.018097324 +0000 UTC m=+117.699999038" watchObservedRunningTime="2026-03-11 13:24:30.022854783 +0000 UTC m=+117.704756497" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.033580 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" event={"ID":"08068104-34d8-440d-8f98-f9a6216af858","Type":"ContainerStarted","Data":"d972361bb06146bbdc4aa442393491cba7ef1d4601e5da3040177b4bf8290720"} Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.033721 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.034608 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.037577 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-mm4gf" podStartSLOduration=65.037564828 podStartE2EDuration="1m5.037564828s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:30.037534497 +0000 UTC m=+117.719436221" watchObservedRunningTime="2026-03-11 13:24:30.037564828 +0000 UTC m=+117.719466542" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.040773 4923 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-hgkc2 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.040827 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" podUID="08068104-34d8-440d-8f98-f9a6216af858" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.048246 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8bhpz" event={"ID":"f4826a3f-d90e-4b7c-a10b-864b78115463","Type":"ContainerStarted","Data":"682b64f43d5a1297e2277885bc3a84d999c6340b1dc5315b2319e15da39528e4"} Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.052717 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4" podStartSLOduration=65.052703283 podStartE2EDuration="1m5.052703283s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:30.051010762 +0000 UTC m=+117.732912476" watchObservedRunningTime="2026-03-11 13:24:30.052703283 +0000 UTC m=+117.734604987" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.071890 4923 generic.go:334] "Generic (PLEG): container finished" podID="1515573d-f360-41f2-854a-4d0fb437af7d" containerID="c2218925671a60600fe4b0a0429a17c8a75dd2f10e71c772a5582d12a781cb9e" exitCode=0 Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.071984 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" event={"ID":"1515573d-f360-41f2-854a-4d0fb437af7d","Type":"ContainerDied","Data":"c2218925671a60600fe4b0a0429a17c8a75dd2f10e71c772a5582d12a781cb9e"} Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.073250 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:30 crc kubenswrapper[4923]: E0311 13:24:30.079392 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:30.579373136 +0000 UTC m=+118.261274850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.082098 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-6jmqc" podStartSLOduration=65.082080703 podStartE2EDuration="1m5.082080703s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:30.08077389 +0000 UTC m=+117.762675604" watchObservedRunningTime="2026-03-11 13:24:30.082080703 +0000 UTC m=+117.763982417" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.090881 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" event={"ID":"b6887dd3-ec66-4364-8eed-f63615c35024","Type":"ContainerStarted","Data":"42bec9be6280f596ed7fc7a0fa6709d2fa9c145343570729298675bbb530a4b9"} Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.090927 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" event={"ID":"b6887dd3-ec66-4364-8eed-f63615c35024","Type":"ContainerStarted","Data":"1fb0655229a7efa4dbd0890f27dedc82b94c6421088752b7e02c607fed31875b"} Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.091677 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.098081 4923 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-vznqg container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.098116 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" podUID="b6887dd3-ec66-4364-8eed-f63615c35024" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.107287 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h7jfj" event={"ID":"7e0c6535-0c45-4d8d-9b97-332888787b98","Type":"ContainerStarted","Data":"9d07c51ae561309a64e9b151fedecd86a9d4005cf48f6693e1c7b863091295ab"} Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.107326 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h7jfj" event={"ID":"7e0c6535-0c45-4d8d-9b97-332888787b98","Type":"ContainerStarted","Data":"64777bacfa4fa7f8f698deadf4302c75cfe3932dd4fa8427e8112158d37241c9"} Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.113066 4923 generic.go:334] "Generic (PLEG): container finished" podID="a67361c9-12cb-4a1a-893c-a735a58da9cd" containerID="456ff3d490e73a68ba921a98a0b330e38b6097ec6e62fde2af541d79452dab2b" exitCode=0 Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.113126 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" event={"ID":"a67361c9-12cb-4a1a-893c-a735a58da9cd","Type":"ContainerDied","Data":"456ff3d490e73a68ba921a98a0b330e38b6097ec6e62fde2af541d79452dab2b"} Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.118862 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dxlk5" event={"ID":"a9d5fee7-2298-4320-a8d1-c1bcb17c4917","Type":"ContainerStarted","Data":"079ad1614cc02f4896470caf2725ad5db81ddbada0b812d8394111b49ca676b9"} Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.118917 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dxlk5" event={"ID":"a9d5fee7-2298-4320-a8d1-c1bcb17c4917","Type":"ContainerStarted","Data":"ba05c6faa23c67747101717245ef33074efb7bdbf20c2055f69caccd5550974e"} Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.141086 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g96sx" event={"ID":"2263dbb9-5bfb-4557-b1f2-2f33ef57db55","Type":"ContainerStarted","Data":"8865bfb80db8db8b3fc1eeb146424d3f9ad4201bf7ad18ecd9645d01c994a0bf"} Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.144050 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-7lx27" event={"ID":"714e692b-982f-4ba9-9d26-8819c83b14fe","Type":"ContainerStarted","Data":"8d9092d9a25a36bb3e0df09c4b8db6332d39b7f51a26d4b1ebeafbc9850f6bca"} Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.144099 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-7lx27" event={"ID":"714e692b-982f-4ba9-9d26-8819c83b14fe","Type":"ContainerStarted","Data":"0a55a5abfb564e1a72921572b65a923587df9a660bdce98cb08793f3cab9d7fc"} Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.144494 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-7lx27" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.152616 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-mgrcg" podStartSLOduration=65.152602193 podStartE2EDuration="1m5.152602193s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:30.126219449 +0000 UTC m=+117.808121163" watchObservedRunningTime="2026-03-11 13:24:30.152602193 +0000 UTC m=+117.834503907" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.167495 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9grj8" event={"ID":"4f20534b-c922-4867-9294-fffaddb02fbf","Type":"ContainerStarted","Data":"4ba11703c0e6197a1b302c72bd81cb1a6308f9a6db9152d3eff5d65b5c1df989"} Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.169656 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9grj8" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.170826 4923 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-9grj8 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.170873 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9grj8" podUID="4f20534b-c922-4867-9294-fffaddb02fbf" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.174235 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:30 crc kubenswrapper[4923]: E0311 13:24:30.184680 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:30.684653729 +0000 UTC m=+118.366555443 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.214004 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l6bf9" podStartSLOduration=65.213984277 podStartE2EDuration="1m5.213984277s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:30.195845517 +0000 UTC m=+117.877747231" watchObservedRunningTime="2026-03-11 13:24:30.213984277 +0000 UTC m=+117.895885981" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.215363 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-8bhpz" podStartSLOduration=65.215359982 podStartE2EDuration="1m5.215359982s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:30.213634688 +0000 UTC m=+117.895536402" watchObservedRunningTime="2026-03-11 13:24:30.215359982 +0000 UTC m=+117.897261696" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.258234 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" podStartSLOduration=65.258219235 podStartE2EDuration="1m5.258219235s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:30.256980294 +0000 UTC m=+117.938882008" watchObservedRunningTime="2026-03-11 13:24:30.258219235 +0000 UTC m=+117.940120949" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.258705 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qkhqm" podStartSLOduration=65.258677517 podStartE2EDuration="1m5.258677517s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:30.229875992 +0000 UTC m=+117.911777706" watchObservedRunningTime="2026-03-11 13:24:30.258677517 +0000 UTC m=+117.940579241" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.282012 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-h7jfj" podStartSLOduration=65.281993985 podStartE2EDuration="1m5.281993985s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:30.276953021 +0000 UTC m=+117.958854735" watchObservedRunningTime="2026-03-11 13:24:30.281993985 +0000 UTC m=+117.963895699" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.286776 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:30 crc kubenswrapper[4923]: E0311 13:24:30.291539 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:30.791526692 +0000 UTC m=+118.473428406 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.338755 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9grj8" podStartSLOduration=65.338726754 podStartE2EDuration="1m5.338726754s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:30.336288383 +0000 UTC m=+118.018190097" watchObservedRunningTime="2026-03-11 13:24:30.338726754 +0000 UTC m=+118.020628458" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.344480 4923 patch_prober.go:28] interesting pod/router-default-5444994796-zd9ns container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 11 13:24:30 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Mar 11 13:24:30 crc kubenswrapper[4923]: [+]process-running ok Mar 11 13:24:30 crc kubenswrapper[4923]: healthz check failed Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.344546 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zd9ns" podUID="3f8730fb-a1a2-440e-8fb2-1bf9134054e2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.362416 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" podStartSLOduration=65.362400321 podStartE2EDuration="1m5.362400321s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:30.360680548 +0000 UTC m=+118.042582262" watchObservedRunningTime="2026-03-11 13:24:30.362400321 +0000 UTC m=+118.044302035" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.385744 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-g96sx" podStartSLOduration=65.38572464 podStartE2EDuration="1m5.38572464s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:30.384638603 +0000 UTC m=+118.066540327" watchObservedRunningTime="2026-03-11 13:24:30.38572464 +0000 UTC m=+118.067626364" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.388307 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:30 crc kubenswrapper[4923]: E0311 13:24:30.388751 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:30.888733785 +0000 UTC m=+118.570635499 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.446942 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dxlk5" podStartSLOduration=65.446925319 podStartE2EDuration="1m5.446925319s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:30.446096909 +0000 UTC m=+118.127998623" watchObservedRunningTime="2026-03-11 13:24:30.446925319 +0000 UTC m=+118.128827033" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.473100 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-7lx27" podStartSLOduration=7.473085049 podStartE2EDuration="7.473085049s" podCreationTimestamp="2026-03-11 13:24:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:30.470951106 +0000 UTC m=+118.152852820" watchObservedRunningTime="2026-03-11 13:24:30.473085049 +0000 UTC m=+118.154986763" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.491750 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:30 crc kubenswrapper[4923]: E0311 13:24:30.492170 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:30.992154922 +0000 UTC m=+118.674056636 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.592722 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:30 crc kubenswrapper[4923]: E0311 13:24:30.592923 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:31.092892923 +0000 UTC m=+118.774794637 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.593560 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:30 crc kubenswrapper[4923]: E0311 13:24:30.593989 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:31.09397138 +0000 UTC m=+118.775873094 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.600452 4923 ???:1] "http: TLS handshake error from 192.168.126.11:38956: no serving certificate available for the kubelet" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.695390 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:30 crc kubenswrapper[4923]: E0311 13:24:30.695619 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:31.195574392 +0000 UTC m=+118.877476106 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.695795 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:30 crc kubenswrapper[4923]: E0311 13:24:30.696202 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:31.196190287 +0000 UTC m=+118.878092001 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.703731 4923 ???:1] "http: TLS handshake error from 192.168.126.11:38964: no serving certificate available for the kubelet" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.796945 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:30 crc kubenswrapper[4923]: E0311 13:24:30.797364 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:31.297331788 +0000 UTC m=+118.979233502 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.803832 4923 ???:1] "http: TLS handshake error from 192.168.126.11:38970: no serving certificate available for the kubelet" Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.898515 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:30 crc kubenswrapper[4923]: E0311 13:24:30.899000 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:31.398983822 +0000 UTC m=+119.080885536 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:30 crc kubenswrapper[4923]: I0311 13:24:30.913328 4923 ???:1] "http: TLS handshake error from 192.168.126.11:38974: no serving certificate available for the kubelet" Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.000422 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:31 crc kubenswrapper[4923]: E0311 13:24:31.000659 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:31.500615534 +0000 UTC m=+119.182517248 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.000837 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:31 crc kubenswrapper[4923]: E0311 13:24:31.001194 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:31.501185658 +0000 UTC m=+119.183087372 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.006012 4923 ???:1] "http: TLS handshake error from 192.168.126.11:38982: no serving certificate available for the kubelet" Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.022701 4923 scope.go:117] "RemoveContainer" containerID="9a9f78b1ff5dfa00fd3d80bafc5c131e5fd12de6d0556229087316aad81df459" Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.064923 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-57kdl"] Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.102364 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:31 crc kubenswrapper[4923]: E0311 13:24:31.102487 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:31.602463482 +0000 UTC m=+119.284365196 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.102877 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:31 crc kubenswrapper[4923]: E0311 13:24:31.103404 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:31.603380796 +0000 UTC m=+119.285282510 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.132559 4923 ???:1] "http: TLS handshake error from 192.168.126.11:38986: no serving certificate available for the kubelet" Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.172380 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" event={"ID":"1515573d-f360-41f2-854a-4d0fb437af7d","Type":"ContainerStarted","Data":"b243f5813417424704007e8c6b5c1b2e0d94ab36de0043048c758d639339f0a7"} Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.172453 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" event={"ID":"1515573d-f360-41f2-854a-4d0fb437af7d","Type":"ContainerStarted","Data":"46dcb19371c68934e74c955e13393adde4ecfe738924cc9a50e0f4c919c3f2eb"} Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.176862 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-dqr7s" event={"ID":"a7bc92e0-e42d-46f4-bc56-5123cb6d36f0","Type":"ContainerStarted","Data":"8f1ce997950bfd486a6ef0b184b41ea3c9acebdc9559a59e3f06211cedd1f347"} Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.185616 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" event={"ID":"a67361c9-12cb-4a1a-893c-a735a58da9cd","Type":"ContainerStarted","Data":"c16cfa5924b7ae5f415e17c2aed98d3ee3f422d6d31c5dcb270791ea865ac48e"} Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.189097 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-qkhqm" event={"ID":"5a3187cd-89a0-4c86-acd9-e325360446eb","Type":"ContainerStarted","Data":"dd698749bde351345548ee514257937f07821526d3573c1bcdd92f88a24a0b04"} Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.191127 4923 generic.go:334] "Generic (PLEG): container finished" podID="122f383b-c9fa-4ab3-8347-e75d16a0189b" containerID="518ed8856b0f90fde2151e77d68426004300ca1a92498bd44fae594d2cecc338" exitCode=0 Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.191178 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29553915-5tqf9" event={"ID":"122f383b-c9fa-4ab3-8347-e75d16a0189b","Type":"ContainerDied","Data":"518ed8856b0f90fde2151e77d68426004300ca1a92498bd44fae594d2cecc338"} Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.192888 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-2c4h7" event={"ID":"0bfac15f-c184-44be-8490-81dfe546f94f","Type":"ContainerStarted","Data":"72e0037d9565d3a412e38017d301a9db62ab811e696df6a023ba14ac54ab12f6"} Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.199973 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4nx54" event={"ID":"ae7e45f7-8c08-435b-990a-5149da33010a","Type":"ContainerStarted","Data":"8cc1ed7bd757dedce233670470a45a57a1ae51f5e95959bcdf51ac94c37ac7a8"} Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.200116 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4nx54" event={"ID":"ae7e45f7-8c08-435b-990a-5149da33010a","Type":"ContainerStarted","Data":"b9b092ca36c6f0cc0869f47bc1a50b4d5ad8851b3208382c20369385ba24713a"} Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.200200 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4nx54" event={"ID":"ae7e45f7-8c08-435b-990a-5149da33010a","Type":"ContainerStarted","Data":"5fa29b634251e6eb13b69c938bead64e42cd926d829edc241ffe061ec39b5b81"} Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.203518 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:31 crc kubenswrapper[4923]: E0311 13:24:31.203779 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:31.703756227 +0000 UTC m=+119.385657941 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.204150 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:31 crc kubenswrapper[4923]: E0311 13:24:31.204961 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:31.704952287 +0000 UTC m=+119.386854001 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.207584 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" podStartSLOduration=66.207571342 podStartE2EDuration="1m6.207571342s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:31.205398548 +0000 UTC m=+118.887300262" watchObservedRunningTime="2026-03-11 13:24:31.207571342 +0000 UTC m=+118.889473056" Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.211136 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lb2kf" event={"ID":"72930709-eec8-4c50-a0da-175e7cbea561","Type":"ContainerStarted","Data":"bfda5438eef6c7d8a58ec44baf534905e12296e6822804e68b01fc7e13e01577"} Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.224505 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h59nb" event={"ID":"717728c1-d998-4fc7-a0d2-ae90e3da4d43","Type":"ContainerStarted","Data":"779b0676a998a01f402f6d5349fdd168a5e116c2c5068d92cda30a573f1a8876"} Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.234289 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" event={"ID":"b0a9d170-2b10-48f1-81ab-898a7df26740","Type":"ContainerStarted","Data":"753bb62b390e212a01737108c17311de28295d98d8cbff51ac1b07c362a88173"} Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.235803 4923 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-vznqg container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.235853 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" podUID="b6887dd3-ec66-4364-8eed-f63615c35024" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.255443 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-9grj8" Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.260783 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5r4gt" Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.296361 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-4nx54" podStartSLOduration=66.296311834 podStartE2EDuration="1m6.296311834s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:31.286779738 +0000 UTC m=+118.968681442" watchObservedRunningTime="2026-03-11 13:24:31.296311834 +0000 UTC m=+118.978213538" Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.305062 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:31 crc kubenswrapper[4923]: E0311 13:24:31.306851 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:31.806825265 +0000 UTC m=+119.488726979 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.361614 4923 patch_prober.go:28] interesting pod/router-default-5444994796-zd9ns container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 11 13:24:31 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Mar 11 13:24:31 crc kubenswrapper[4923]: [+]process-running ok Mar 11 13:24:31 crc kubenswrapper[4923]: healthz check failed Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.361687 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zd9ns" podUID="3f8730fb-a1a2-440e-8fb2-1bf9134054e2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.408797 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:31 crc kubenswrapper[4923]: E0311 13:24:31.409203 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:31.909177906 +0000 UTC m=+119.591079620 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.419655 4923 ???:1] "http: TLS handshake error from 192.168.126.11:38988: no serving certificate available for the kubelet" Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.445775 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" podStartSLOduration=66.445757184 podStartE2EDuration="1m6.445757184s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:31.441904928 +0000 UTC m=+119.123806642" watchObservedRunningTime="2026-03-11 13:24:31.445757184 +0000 UTC m=+119.127658898" Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.497602 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.510895 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:31 crc kubenswrapper[4923]: E0311 13:24:31.511250 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:32.011231919 +0000 UTC m=+119.693133633 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.612440 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:31 crc kubenswrapper[4923]: E0311 13:24:31.612783 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:32.11276692 +0000 UTC m=+119.794668634 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.638691 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lb2kf" podStartSLOduration=66.638675294 podStartE2EDuration="1m6.638675294s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:31.637482323 +0000 UTC m=+119.319384037" watchObservedRunningTime="2026-03-11 13:24:31.638675294 +0000 UTC m=+119.320577008" Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.713972 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:31 crc kubenswrapper[4923]: E0311 13:24:31.714610 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:32.214569857 +0000 UTC m=+119.896471581 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.815841 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:31 crc kubenswrapper[4923]: E0311 13:24:31.816598 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:32.316573919 +0000 UTC m=+119.998475633 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.836612 4923 ???:1] "http: TLS handshake error from 192.168.126.11:39002: no serving certificate available for the kubelet" Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.879616 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-l6bf9" Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.916680 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:31 crc kubenswrapper[4923]: E0311 13:24:31.916913 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:32.416884539 +0000 UTC m=+120.098786263 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:31 crc kubenswrapper[4923]: I0311 13:24:31.917005 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:31 crc kubenswrapper[4923]: E0311 13:24:31.917290 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:32.417278169 +0000 UTC m=+120.099179883 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.018518 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:32 crc kubenswrapper[4923]: E0311 13:24:32.018685 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:32.518661026 +0000 UTC m=+120.200562740 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.018818 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:32 crc kubenswrapper[4923]: E0311 13:24:32.019168 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:32.519154558 +0000 UTC m=+120.201056272 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.120105 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:32 crc kubenswrapper[4923]: E0311 13:24:32.120249 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:32.620222546 +0000 UTC m=+120.302124250 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.120395 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:32 crc kubenswrapper[4923]: E0311 13:24:32.120671 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:32.620659948 +0000 UTC m=+120.302561662 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.221386 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:32 crc kubenswrapper[4923]: E0311 13:24:32.221583 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:32.721558523 +0000 UTC m=+120.403460237 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.221685 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:32 crc kubenswrapper[4923]: E0311 13:24:32.221956 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:32.721944892 +0000 UTC m=+120.403846606 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.236643 4923 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-lc5r4 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.236709 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4" podUID="81c57b2e-1df9-44c1-9749-21fe5780f479" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.31:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.246123 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" event={"ID":"b0a9d170-2b10-48f1-81ab-898a7df26740","Type":"ContainerStarted","Data":"8572055c1b0fc83f0df1193184d05ff61a6beb3c8037bc7be464b80dffff7802"} Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.261379 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.275271 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c84697dbf2629c6bfcdf82a7ff158126b4c1ec298d635619eff5fbe84af05118"} Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.277679 4923 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-vznqg container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.280056 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" podUID="b6887dd3-ec66-4364-8eed-f63615c35024" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.277791 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" podUID="a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04" containerName="kube-multus-additional-cni-plugins" containerID="cri-o://d348f00e8a6e4b07e7a23195b2d70fdf885db23f4ce1b740a3fcb77ccd7756ec" gracePeriod=30 Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.342521 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.343252 4923 patch_prober.go:28] interesting pod/router-default-5444994796-zd9ns container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 11 13:24:32 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Mar 11 13:24:32 crc kubenswrapper[4923]: [+]process-running ok Mar 11 13:24:32 crc kubenswrapper[4923]: healthz check failed Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.343293 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zd9ns" podUID="3f8730fb-a1a2-440e-8fb2-1bf9134054e2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 11 13:24:32 crc kubenswrapper[4923]: E0311 13:24:32.366271 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:32.866235974 +0000 UTC m=+120.548137688 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.385705 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=16.385686177 podStartE2EDuration="16.385686177s" podCreationTimestamp="2026-03-11 13:24:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:32.319699029 +0000 UTC m=+120.001600743" watchObservedRunningTime="2026-03-11 13:24:32.385686177 +0000 UTC m=+120.067587891" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.394601 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8f4cb"] Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.400808 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8f4cb" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.403617 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8f4cb"] Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.404162 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.462736 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-lc5r4" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.465184 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52f59385-23ad-4163-a766-b9d21e57ad6c-catalog-content\") pod \"certified-operators-8f4cb\" (UID: \"52f59385-23ad-4163-a766-b9d21e57ad6c\") " pod="openshift-marketplace/certified-operators-8f4cb" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.465250 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52f59385-23ad-4163-a766-b9d21e57ad6c-utilities\") pod \"certified-operators-8f4cb\" (UID: \"52f59385-23ad-4163-a766-b9d21e57ad6c\") " pod="openshift-marketplace/certified-operators-8f4cb" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.465291 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.465332 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d48t2\" (UniqueName: \"kubernetes.io/projected/52f59385-23ad-4163-a766-b9d21e57ad6c-kube-api-access-d48t2\") pod \"certified-operators-8f4cb\" (UID: \"52f59385-23ad-4163-a766-b9d21e57ad6c\") " pod="openshift-marketplace/certified-operators-8f4cb" Mar 11 13:24:32 crc kubenswrapper[4923]: E0311 13:24:32.465649 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:32.965637382 +0000 UTC m=+120.647539096 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.541625 4923 ???:1] "http: TLS handshake error from 192.168.126.11:39016: no serving certificate available for the kubelet" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.559475 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cqdq8"] Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.560365 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cqdq8" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.563767 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.567803 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.567936 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52f59385-23ad-4163-a766-b9d21e57ad6c-utilities\") pod \"certified-operators-8f4cb\" (UID: \"52f59385-23ad-4163-a766-b9d21e57ad6c\") " pod="openshift-marketplace/certified-operators-8f4cb" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.568030 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d48t2\" (UniqueName: \"kubernetes.io/projected/52f59385-23ad-4163-a766-b9d21e57ad6c-kube-api-access-d48t2\") pod \"certified-operators-8f4cb\" (UID: \"52f59385-23ad-4163-a766-b9d21e57ad6c\") " pod="openshift-marketplace/certified-operators-8f4cb" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.568060 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52f59385-23ad-4163-a766-b9d21e57ad6c-catalog-content\") pod \"certified-operators-8f4cb\" (UID: \"52f59385-23ad-4163-a766-b9d21e57ad6c\") " pod="openshift-marketplace/certified-operators-8f4cb" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.568553 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52f59385-23ad-4163-a766-b9d21e57ad6c-catalog-content\") pod \"certified-operators-8f4cb\" (UID: \"52f59385-23ad-4163-a766-b9d21e57ad6c\") " pod="openshift-marketplace/certified-operators-8f4cb" Mar 11 13:24:32 crc kubenswrapper[4923]: E0311 13:24:32.568617 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:33.068603458 +0000 UTC m=+120.750505162 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.568812 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52f59385-23ad-4163-a766-b9d21e57ad6c-utilities\") pod \"certified-operators-8f4cb\" (UID: \"52f59385-23ad-4163-a766-b9d21e57ad6c\") " pod="openshift-marketplace/certified-operators-8f4cb" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.586497 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cqdq8"] Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.629427 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d48t2\" (UniqueName: \"kubernetes.io/projected/52f59385-23ad-4163-a766-b9d21e57ad6c-kube-api-access-d48t2\") pod \"certified-operators-8f4cb\" (UID: \"52f59385-23ad-4163-a766-b9d21e57ad6c\") " pod="openshift-marketplace/certified-operators-8f4cb" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.646168 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-g25ck"] Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.652530 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" podUID="16d21cac-dc41-43fb-a391-8864085617ed" containerName="controller-manager" containerID="cri-o://bff21e897d1aa1ece817f88e76ea011624e18cc06b8cb1806c01ebcf4ddcb28e" gracePeriod=30 Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.669129 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mzqq\" (UniqueName: \"kubernetes.io/projected/7ce0bb4a-0841-4449-bc02-6c219731272e-kube-api-access-8mzqq\") pod \"community-operators-cqdq8\" (UID: \"7ce0bb4a-0841-4449-bc02-6c219731272e\") " pod="openshift-marketplace/community-operators-cqdq8" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.669225 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.669270 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ce0bb4a-0841-4449-bc02-6c219731272e-utilities\") pod \"community-operators-cqdq8\" (UID: \"7ce0bb4a-0841-4449-bc02-6c219731272e\") " pod="openshift-marketplace/community-operators-cqdq8" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.669290 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ce0bb4a-0841-4449-bc02-6c219731272e-catalog-content\") pod \"community-operators-cqdq8\" (UID: \"7ce0bb4a-0841-4449-bc02-6c219731272e\") " pod="openshift-marketplace/community-operators-cqdq8" Mar 11 13:24:32 crc kubenswrapper[4923]: E0311 13:24:32.669570 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:33.169558544 +0000 UTC m=+120.851460258 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.699837 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql"] Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.700084 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" podUID="cfda8f36-49b9-4081-bbea-f311dd0a87a1" containerName="route-controller-manager" containerID="cri-o://a23af8b5446190635a930538e80a26c355ca590fd3665b3e9da27144b5ad9381" gracePeriod=30 Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.745139 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8f4cb" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.771885 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.772089 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ce0bb4a-0841-4449-bc02-6c219731272e-catalog-content\") pod \"community-operators-cqdq8\" (UID: \"7ce0bb4a-0841-4449-bc02-6c219731272e\") " pod="openshift-marketplace/community-operators-cqdq8" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.772134 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mzqq\" (UniqueName: \"kubernetes.io/projected/7ce0bb4a-0841-4449-bc02-6c219731272e-kube-api-access-8mzqq\") pod \"community-operators-cqdq8\" (UID: \"7ce0bb4a-0841-4449-bc02-6c219731272e\") " pod="openshift-marketplace/community-operators-cqdq8" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.772197 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ce0bb4a-0841-4449-bc02-6c219731272e-utilities\") pod \"community-operators-cqdq8\" (UID: \"7ce0bb4a-0841-4449-bc02-6c219731272e\") " pod="openshift-marketplace/community-operators-cqdq8" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.772575 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ce0bb4a-0841-4449-bc02-6c219731272e-utilities\") pod \"community-operators-cqdq8\" (UID: \"7ce0bb4a-0841-4449-bc02-6c219731272e\") " pod="openshift-marketplace/community-operators-cqdq8" Mar 11 13:24:32 crc kubenswrapper[4923]: E0311 13:24:32.772644 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:33.272629782 +0000 UTC m=+120.954531496 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.772832 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ce0bb4a-0841-4449-bc02-6c219731272e-catalog-content\") pod \"community-operators-cqdq8\" (UID: \"7ce0bb4a-0841-4449-bc02-6c219731272e\") " pod="openshift-marketplace/community-operators-cqdq8" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.804419 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mzqq\" (UniqueName: \"kubernetes.io/projected/7ce0bb4a-0841-4449-bc02-6c219731272e-kube-api-access-8mzqq\") pod \"community-operators-cqdq8\" (UID: \"7ce0bb4a-0841-4449-bc02-6c219731272e\") " pod="openshift-marketplace/community-operators-cqdq8" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.810806 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rdvss"] Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.828005 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rdvss" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.853228 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rdvss"] Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.876283 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:32 crc kubenswrapper[4923]: E0311 13:24:32.876682 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:33.376669905 +0000 UTC m=+121.058571619 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.888763 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cqdq8" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.918785 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29553915-5tqf9" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.977126 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:32 crc kubenswrapper[4923]: E0311 13:24:32.977445 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:33.477424576 +0000 UTC m=+121.159326290 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.977689 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bdd5510-3bb1-42f0-8646-c45f11bdeafe-catalog-content\") pod \"certified-operators-rdvss\" (UID: \"8bdd5510-3bb1-42f0-8646-c45f11bdeafe\") " pod="openshift-marketplace/certified-operators-rdvss" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.977731 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bdd5510-3bb1-42f0-8646-c45f11bdeafe-utilities\") pod \"certified-operators-rdvss\" (UID: \"8bdd5510-3bb1-42f0-8646-c45f11bdeafe\") " pod="openshift-marketplace/certified-operators-rdvss" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.977768 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:32 crc kubenswrapper[4923]: I0311 13:24:32.977799 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drk5r\" (UniqueName: \"kubernetes.io/projected/8bdd5510-3bb1-42f0-8646-c45f11bdeafe-kube-api-access-drk5r\") pod \"certified-operators-rdvss\" (UID: \"8bdd5510-3bb1-42f0-8646-c45f11bdeafe\") " pod="openshift-marketplace/certified-operators-rdvss" Mar 11 13:24:32 crc kubenswrapper[4923]: E0311 13:24:32.978061 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:33.478049882 +0000 UTC m=+121.159951596 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.030067 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-s9ncx"] Mar 11 13:24:33 crc kubenswrapper[4923]: E0311 13:24:33.030433 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="122f383b-c9fa-4ab3-8347-e75d16a0189b" containerName="collect-profiles" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.030449 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="122f383b-c9fa-4ab3-8347-e75d16a0189b" containerName="collect-profiles" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.030618 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="122f383b-c9fa-4ab3-8347-e75d16a0189b" containerName="collect-profiles" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.031734 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s9ncx" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.055385 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s9ncx"] Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.079037 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/122f383b-c9fa-4ab3-8347-e75d16a0189b-config-volume\") pod \"122f383b-c9fa-4ab3-8347-e75d16a0189b\" (UID: \"122f383b-c9fa-4ab3-8347-e75d16a0189b\") " Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.079127 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/122f383b-c9fa-4ab3-8347-e75d16a0189b-secret-volume\") pod \"122f383b-c9fa-4ab3-8347-e75d16a0189b\" (UID: \"122f383b-c9fa-4ab3-8347-e75d16a0189b\") " Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.079303 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.079335 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4r7tg\" (UniqueName: \"kubernetes.io/projected/122f383b-c9fa-4ab3-8347-e75d16a0189b-kube-api-access-4r7tg\") pod \"122f383b-c9fa-4ab3-8347-e75d16a0189b\" (UID: \"122f383b-c9fa-4ab3-8347-e75d16a0189b\") " Mar 11 13:24:33 crc kubenswrapper[4923]: E0311 13:24:33.081492 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:33.581459319 +0000 UTC m=+121.263361033 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.081536 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.081598 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbba7a9b-2c9b-4770-ac96-b89390457eab-utilities\") pod \"community-operators-s9ncx\" (UID: \"dbba7a9b-2c9b-4770-ac96-b89390457eab\") " pod="openshift-marketplace/community-operators-s9ncx" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.081635 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbba7a9b-2c9b-4770-ac96-b89390457eab-catalog-content\") pod \"community-operators-s9ncx\" (UID: \"dbba7a9b-2c9b-4770-ac96-b89390457eab\") " pod="openshift-marketplace/community-operators-s9ncx" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.081708 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drk5r\" (UniqueName: \"kubernetes.io/projected/8bdd5510-3bb1-42f0-8646-c45f11bdeafe-kube-api-access-drk5r\") pod \"certified-operators-rdvss\" (UID: \"8bdd5510-3bb1-42f0-8646-c45f11bdeafe\") " pod="openshift-marketplace/certified-operators-rdvss" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.081803 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bdd5510-3bb1-42f0-8646-c45f11bdeafe-catalog-content\") pod \"certified-operators-rdvss\" (UID: \"8bdd5510-3bb1-42f0-8646-c45f11bdeafe\") " pod="openshift-marketplace/certified-operators-rdvss" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.081872 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w48wt\" (UniqueName: \"kubernetes.io/projected/dbba7a9b-2c9b-4770-ac96-b89390457eab-kube-api-access-w48wt\") pod \"community-operators-s9ncx\" (UID: \"dbba7a9b-2c9b-4770-ac96-b89390457eab\") " pod="openshift-marketplace/community-operators-s9ncx" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.081913 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bdd5510-3bb1-42f0-8646-c45f11bdeafe-utilities\") pod \"certified-operators-rdvss\" (UID: \"8bdd5510-3bb1-42f0-8646-c45f11bdeafe\") " pod="openshift-marketplace/certified-operators-rdvss" Mar 11 13:24:33 crc kubenswrapper[4923]: E0311 13:24:33.082275 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:33.582252158 +0000 UTC m=+121.264153872 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.082413 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bdd5510-3bb1-42f0-8646-c45f11bdeafe-utilities\") pod \"certified-operators-rdvss\" (UID: \"8bdd5510-3bb1-42f0-8646-c45f11bdeafe\") " pod="openshift-marketplace/certified-operators-rdvss" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.082928 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/122f383b-c9fa-4ab3-8347-e75d16a0189b-config-volume" (OuterVolumeSpecName: "config-volume") pod "122f383b-c9fa-4ab3-8347-e75d16a0189b" (UID: "122f383b-c9fa-4ab3-8347-e75d16a0189b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.083181 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bdd5510-3bb1-42f0-8646-c45f11bdeafe-catalog-content\") pod \"certified-operators-rdvss\" (UID: \"8bdd5510-3bb1-42f0-8646-c45f11bdeafe\") " pod="openshift-marketplace/certified-operators-rdvss" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.126142 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/122f383b-c9fa-4ab3-8347-e75d16a0189b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "122f383b-c9fa-4ab3-8347-e75d16a0189b" (UID: "122f383b-c9fa-4ab3-8347-e75d16a0189b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.127121 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/122f383b-c9fa-4ab3-8347-e75d16a0189b-kube-api-access-4r7tg" (OuterVolumeSpecName: "kube-api-access-4r7tg") pod "122f383b-c9fa-4ab3-8347-e75d16a0189b" (UID: "122f383b-c9fa-4ab3-8347-e75d16a0189b"). InnerVolumeSpecName "kube-api-access-4r7tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.151388 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drk5r\" (UniqueName: \"kubernetes.io/projected/8bdd5510-3bb1-42f0-8646-c45f11bdeafe-kube-api-access-drk5r\") pod \"certified-operators-rdvss\" (UID: \"8bdd5510-3bb1-42f0-8646-c45f11bdeafe\") " pod="openshift-marketplace/certified-operators-rdvss" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.183704 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.185302 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w48wt\" (UniqueName: \"kubernetes.io/projected/dbba7a9b-2c9b-4770-ac96-b89390457eab-kube-api-access-w48wt\") pod \"community-operators-s9ncx\" (UID: \"dbba7a9b-2c9b-4770-ac96-b89390457eab\") " pod="openshift-marketplace/community-operators-s9ncx" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.185383 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbba7a9b-2c9b-4770-ac96-b89390457eab-utilities\") pod \"community-operators-s9ncx\" (UID: \"dbba7a9b-2c9b-4770-ac96-b89390457eab\") " pod="openshift-marketplace/community-operators-s9ncx" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.185402 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbba7a9b-2c9b-4770-ac96-b89390457eab-catalog-content\") pod \"community-operators-s9ncx\" (UID: \"dbba7a9b-2c9b-4770-ac96-b89390457eab\") " pod="openshift-marketplace/community-operators-s9ncx" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.185471 4923 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/122f383b-c9fa-4ab3-8347-e75d16a0189b-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.185481 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4r7tg\" (UniqueName: \"kubernetes.io/projected/122f383b-c9fa-4ab3-8347-e75d16a0189b-kube-api-access-4r7tg\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.185490 4923 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/122f383b-c9fa-4ab3-8347-e75d16a0189b-config-volume\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.187475 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbba7a9b-2c9b-4770-ac96-b89390457eab-catalog-content\") pod \"community-operators-s9ncx\" (UID: \"dbba7a9b-2c9b-4770-ac96-b89390457eab\") " pod="openshift-marketplace/community-operators-s9ncx" Mar 11 13:24:33 crc kubenswrapper[4923]: E0311 13:24:33.187539 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:33.687523711 +0000 UTC m=+121.369425425 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.187987 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbba7a9b-2c9b-4770-ac96-b89390457eab-utilities\") pod \"community-operators-s9ncx\" (UID: \"dbba7a9b-2c9b-4770-ac96-b89390457eab\") " pod="openshift-marketplace/community-operators-s9ncx" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.190517 4923 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.227755 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rdvss" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.248564 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w48wt\" (UniqueName: \"kubernetes.io/projected/dbba7a9b-2c9b-4770-ac96-b89390457eab-kube-api-access-w48wt\") pod \"community-operators-s9ncx\" (UID: \"dbba7a9b-2c9b-4770-ac96-b89390457eab\") " pod="openshift-marketplace/community-operators-s9ncx" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.287282 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:33 crc kubenswrapper[4923]: E0311 13:24:33.287767 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:33.787743449 +0000 UTC m=+121.469645163 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.317179 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" event={"ID":"b0a9d170-2b10-48f1-81ab-898a7df26740","Type":"ContainerStarted","Data":"84e93cd18c0de5992b58c6bf6fa5b0d2ec1e32fc4fc0291b3ff75f6856788e99"} Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.319985 4923 generic.go:334] "Generic (PLEG): container finished" podID="cfda8f36-49b9-4081-bbea-f311dd0a87a1" containerID="a23af8b5446190635a930538e80a26c355ca590fd3665b3e9da27144b5ad9381" exitCode=0 Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.320040 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" event={"ID":"cfda8f36-49b9-4081-bbea-f311dd0a87a1","Type":"ContainerDied","Data":"a23af8b5446190635a930538e80a26c355ca590fd3665b3e9da27144b5ad9381"} Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.331729 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29553915-5tqf9" event={"ID":"122f383b-c9fa-4ab3-8347-e75d16a0189b","Type":"ContainerDied","Data":"9cc6f13e7b62558bf379c1647be8887c5fa2a19760e4486387981eb7d4784395"} Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.331794 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9cc6f13e7b62558bf379c1647be8887c5fa2a19760e4486387981eb7d4784395" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.331980 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29553915-5tqf9" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.345198 4923 generic.go:334] "Generic (PLEG): container finished" podID="16d21cac-dc41-43fb-a391-8864085617ed" containerID="bff21e897d1aa1ece817f88e76ea011624e18cc06b8cb1806c01ebcf4ddcb28e" exitCode=0 Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.347822 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" event={"ID":"16d21cac-dc41-43fb-a391-8864085617ed","Type":"ContainerDied","Data":"bff21e897d1aa1ece817f88e76ea011624e18cc06b8cb1806c01ebcf4ddcb28e"} Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.365089 4923 patch_prober.go:28] interesting pod/router-default-5444994796-zd9ns container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 11 13:24:33 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Mar 11 13:24:33 crc kubenswrapper[4923]: [+]process-running ok Mar 11 13:24:33 crc kubenswrapper[4923]: healthz check failed Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.365151 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zd9ns" podUID="3f8730fb-a1a2-440e-8fb2-1bf9134054e2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.390253 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:33 crc kubenswrapper[4923]: E0311 13:24:33.391884 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:33.891860013 +0000 UTC m=+121.573761727 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.440030 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.451703 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s9ncx" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.491051 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.491221 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcgv5\" (UniqueName: \"kubernetes.io/projected/cfda8f36-49b9-4081-bbea-f311dd0a87a1-kube-api-access-mcgv5\") pod \"cfda8f36-49b9-4081-bbea-f311dd0a87a1\" (UID: \"cfda8f36-49b9-4081-bbea-f311dd0a87a1\") " Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.491555 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cfda8f36-49b9-4081-bbea-f311dd0a87a1-client-ca\") pod \"cfda8f36-49b9-4081-bbea-f311dd0a87a1\" (UID: \"cfda8f36-49b9-4081-bbea-f311dd0a87a1\") " Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.491605 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfda8f36-49b9-4081-bbea-f311dd0a87a1-config\") pod \"cfda8f36-49b9-4081-bbea-f311dd0a87a1\" (UID: \"cfda8f36-49b9-4081-bbea-f311dd0a87a1\") " Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.491724 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cfda8f36-49b9-4081-bbea-f311dd0a87a1-serving-cert\") pod \"cfda8f36-49b9-4081-bbea-f311dd0a87a1\" (UID: \"cfda8f36-49b9-4081-bbea-f311dd0a87a1\") " Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.492056 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:33 crc kubenswrapper[4923]: E0311 13:24:33.492314 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:33.992303327 +0000 UTC m=+121.674205041 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.492467 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfda8f36-49b9-4081-bbea-f311dd0a87a1-config" (OuterVolumeSpecName: "config") pod "cfda8f36-49b9-4081-bbea-f311dd0a87a1" (UID: "cfda8f36-49b9-4081-bbea-f311dd0a87a1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.492633 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfda8f36-49b9-4081-bbea-f311dd0a87a1-client-ca" (OuterVolumeSpecName: "client-ca") pod "cfda8f36-49b9-4081-bbea-f311dd0a87a1" (UID: "cfda8f36-49b9-4081-bbea-f311dd0a87a1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.496148 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfda8f36-49b9-4081-bbea-f311dd0a87a1-kube-api-access-mcgv5" (OuterVolumeSpecName: "kube-api-access-mcgv5") pod "cfda8f36-49b9-4081-bbea-f311dd0a87a1" (UID: "cfda8f36-49b9-4081-bbea-f311dd0a87a1"). InnerVolumeSpecName "kube-api-access-mcgv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:33 crc kubenswrapper[4923]: E0311 13:24:33.501723 4923 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod534d6af5_be40_40a6_accb_8b8b11cbf774.slice/crio-5a7f61785480d84b2c1b490d19c6dd1f7d6f0145972e95a1a0567fe402d50576\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod122f383b_c9fa_4ab3_8347_e75d16a0189b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddbba7a9b_2c9b_4770_ac96_b89390457eab.slice\": RecentStats: unable to find data in memory cache]" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.510400 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfda8f36-49b9-4081-bbea-f311dd0a87a1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "cfda8f36-49b9-4081-bbea-f311dd0a87a1" (UID: "cfda8f36-49b9-4081-bbea-f311dd0a87a1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.517140 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8f4cb"] Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.537977 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cqdq8"] Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.593786 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/16d21cac-dc41-43fb-a391-8864085617ed-proxy-ca-bundles\") pod \"16d21cac-dc41-43fb-a391-8864085617ed\" (UID: \"16d21cac-dc41-43fb-a391-8864085617ed\") " Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.593980 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.594010 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16d21cac-dc41-43fb-a391-8864085617ed-config\") pod \"16d21cac-dc41-43fb-a391-8864085617ed\" (UID: \"16d21cac-dc41-43fb-a391-8864085617ed\") " Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.594127 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16d21cac-dc41-43fb-a391-8864085617ed-serving-cert\") pod \"16d21cac-dc41-43fb-a391-8864085617ed\" (UID: \"16d21cac-dc41-43fb-a391-8864085617ed\") " Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.594178 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r582n\" (UniqueName: \"kubernetes.io/projected/16d21cac-dc41-43fb-a391-8864085617ed-kube-api-access-r582n\") pod \"16d21cac-dc41-43fb-a391-8864085617ed\" (UID: \"16d21cac-dc41-43fb-a391-8864085617ed\") " Mar 11 13:24:33 crc kubenswrapper[4923]: E0311 13:24:33.594210 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:34.094174726 +0000 UTC m=+121.776076490 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.594264 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/16d21cac-dc41-43fb-a391-8864085617ed-client-ca\") pod \"16d21cac-dc41-43fb-a391-8864085617ed\" (UID: \"16d21cac-dc41-43fb-a391-8864085617ed\") " Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.594759 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16d21cac-dc41-43fb-a391-8864085617ed-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "16d21cac-dc41-43fb-a391-8864085617ed" (UID: "16d21cac-dc41-43fb-a391-8864085617ed"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.594901 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.595068 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcgv5\" (UniqueName: \"kubernetes.io/projected/cfda8f36-49b9-4081-bbea-f311dd0a87a1-kube-api-access-mcgv5\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.595079 4923 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cfda8f36-49b9-4081-bbea-f311dd0a87a1-client-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.595087 4923 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/16d21cac-dc41-43fb-a391-8864085617ed-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.595096 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfda8f36-49b9-4081-bbea-f311dd0a87a1-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.595105 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cfda8f36-49b9-4081-bbea-f311dd0a87a1-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:33 crc kubenswrapper[4923]: E0311 13:24:33.595378 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:34.095371496 +0000 UTC m=+121.777273210 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.595480 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16d21cac-dc41-43fb-a391-8864085617ed-config" (OuterVolumeSpecName: "config") pod "16d21cac-dc41-43fb-a391-8864085617ed" (UID: "16d21cac-dc41-43fb-a391-8864085617ed"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.595743 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16d21cac-dc41-43fb-a391-8864085617ed-client-ca" (OuterVolumeSpecName: "client-ca") pod "16d21cac-dc41-43fb-a391-8864085617ed" (UID: "16d21cac-dc41-43fb-a391-8864085617ed"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.601702 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16d21cac-dc41-43fb-a391-8864085617ed-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "16d21cac-dc41-43fb-a391-8864085617ed" (UID: "16d21cac-dc41-43fb-a391-8864085617ed"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.602801 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rdvss"] Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.609287 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16d21cac-dc41-43fb-a391-8864085617ed-kube-api-access-r582n" (OuterVolumeSpecName: "kube-api-access-r582n") pod "16d21cac-dc41-43fb-a391-8864085617ed" (UID: "16d21cac-dc41-43fb-a391-8864085617ed"). InnerVolumeSpecName "kube-api-access-r582n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.695805 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:33 crc kubenswrapper[4923]: E0311 13:24:33.695917 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:34.195892111 +0000 UTC m=+121.877793825 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.695985 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.696022 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.696077 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16d21cac-dc41-43fb-a391-8864085617ed-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.696097 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16d21cac-dc41-43fb-a391-8864085617ed-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.696106 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r582n\" (UniqueName: \"kubernetes.io/projected/16d21cac-dc41-43fb-a391-8864085617ed-kube-api-access-r582n\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.696115 4923 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/16d21cac-dc41-43fb-a391-8864085617ed-client-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:33 crc kubenswrapper[4923]: E0311 13:24:33.696355 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:34.196329182 +0000 UTC m=+121.878230896 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.734481 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s9ncx"] Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.796933 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:33 crc kubenswrapper[4923]: E0311 13:24:33.797157 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:34.297113724 +0000 UTC m=+121.979015438 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.797446 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:33 crc kubenswrapper[4923]: E0311 13:24:33.797732 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:34.297723879 +0000 UTC m=+121.979625583 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:33 crc kubenswrapper[4923]: W0311 13:24:33.845037 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddbba7a9b_2c9b_4770_ac96_b89390457eab.slice/crio-184cf4856a6dd957f402dab4aa900b8d61954d494dad4611ca54a419a5cf33f1 WatchSource:0}: Error finding container 184cf4856a6dd957f402dab4aa900b8d61954d494dad4611ca54a419a5cf33f1: Status 404 returned error can't find the container with id 184cf4856a6dd957f402dab4aa900b8d61954d494dad4611ca54a419a5cf33f1 Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.870799 4923 ???:1] "http: TLS handshake error from 192.168.126.11:39024: no serving certificate available for the kubelet" Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.899161 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:33 crc kubenswrapper[4923]: E0311 13:24:33.899458 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:34.399424374 +0000 UTC m=+122.081326108 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:33 crc kubenswrapper[4923]: I0311 13:24:33.899720 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:33 crc kubenswrapper[4923]: E0311 13:24:33.900136 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:34.400121981 +0000 UTC m=+122.082023715 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.000904 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:34 crc kubenswrapper[4923]: E0311 13:24:34.001001 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:34.500967444 +0000 UTC m=+122.182869158 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.001123 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:34 crc kubenswrapper[4923]: E0311 13:24:34.001440 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:34.501417195 +0000 UTC m=+122.183318899 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.102538 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:34 crc kubenswrapper[4923]: E0311 13:24:34.102699 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-11 13:24:34.602675069 +0000 UTC m=+122.284576793 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.102753 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:34 crc kubenswrapper[4923]: E0311 13:24:34.103009 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-11 13:24:34.603001047 +0000 UTC m=+122.284902761 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fwfgr" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.173566 4923 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-11T13:24:33.190533027Z","Handler":null,"Name":""} Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.176275 4923 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.176332 4923 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.195847 4923 patch_prober.go:28] interesting pod/downloads-7954f5f757-b5jmc container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" start-of-body= Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.195902 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-b5jmc" podUID="c7215b86-4f1e-44fb-b0a9-c7d261ce4a54" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.195856 4923 patch_prober.go:28] interesting pod/downloads-7954f5f757-b5jmc container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" start-of-body= Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.196050 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-b5jmc" podUID="c7215b86-4f1e-44fb-b0a9-c7d261ce4a54" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.5:8080/\": dial tcp 10.217.0.5:8080: connect: connection refused" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.204061 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.217248 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.277181 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5"] Mar 11 13:24:34 crc kubenswrapper[4923]: E0311 13:24:34.278001 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfda8f36-49b9-4081-bbea-f311dd0a87a1" containerName="route-controller-manager" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.278020 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfda8f36-49b9-4081-bbea-f311dd0a87a1" containerName="route-controller-manager" Mar 11 13:24:34 crc kubenswrapper[4923]: E0311 13:24:34.278036 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16d21cac-dc41-43fb-a391-8864085617ed" containerName="controller-manager" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.278044 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="16d21cac-dc41-43fb-a391-8864085617ed" containerName="controller-manager" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.278159 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfda8f36-49b9-4081-bbea-f311dd0a87a1" containerName="route-controller-manager" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.278174 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="16d21cac-dc41-43fb-a391-8864085617ed" containerName="controller-manager" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.278625 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.293245 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5"] Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.305371 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.317652 4923 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.317722 4923 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.337471 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fwfgr\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.343599 4923 patch_prober.go:28] interesting pod/router-default-5444994796-zd9ns container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 11 13:24:34 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Mar 11 13:24:34 crc kubenswrapper[4923]: [+]process-running ok Mar 11 13:24:34 crc kubenswrapper[4923]: healthz check failed Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.343655 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zd9ns" podUID="3f8730fb-a1a2-440e-8fb2-1bf9134054e2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.353450 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.355614 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-g25ck" event={"ID":"16d21cac-dc41-43fb-a391-8864085617ed","Type":"ContainerDied","Data":"f339e8b021e35c253199ac83505e1957927c26e46bb1b6044f6e75b9233de1ae"} Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.355682 4923 scope.go:117] "RemoveContainer" containerID="bff21e897d1aa1ece817f88e76ea011624e18cc06b8cb1806c01ebcf4ddcb28e" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.359333 4923 generic.go:334] "Generic (PLEG): container finished" podID="52f59385-23ad-4163-a766-b9d21e57ad6c" containerID="c573b74eea83b701b8f005938e15c3d4ec63a4098b8bdd76d2ba418a747883cd" exitCode=0 Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.359436 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8f4cb" event={"ID":"52f59385-23ad-4163-a766-b9d21e57ad6c","Type":"ContainerDied","Data":"c573b74eea83b701b8f005938e15c3d4ec63a4098b8bdd76d2ba418a747883cd"} Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.359458 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8f4cb" event={"ID":"52f59385-23ad-4163-a766-b9d21e57ad6c","Type":"ContainerStarted","Data":"dcf2b128619685189cc795ffc1f620befcc350a4ec4d6de3e7a6de4148822861"} Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.363165 4923 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.367569 4923 generic.go:334] "Generic (PLEG): container finished" podID="8bdd5510-3bb1-42f0-8646-c45f11bdeafe" containerID="e73ae67b9317111c1da0092fec6fb392e4d3eb633bc285917b7dd84798ca2a57" exitCode=0 Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.367653 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdvss" event={"ID":"8bdd5510-3bb1-42f0-8646-c45f11bdeafe","Type":"ContainerDied","Data":"e73ae67b9317111c1da0092fec6fb392e4d3eb633bc285917b7dd84798ca2a57"} Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.367681 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdvss" event={"ID":"8bdd5510-3bb1-42f0-8646-c45f11bdeafe","Type":"ContainerStarted","Data":"28506726bb33e4937d4c58ccc316a19ff456adddcd5a84b46308d237de1b7cdf"} Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.371360 4923 generic.go:334] "Generic (PLEG): container finished" podID="dbba7a9b-2c9b-4770-ac96-b89390457eab" containerID="cf4922608027344a703c1ef9f815054a3bdba81a21392a04abf1e2d76d3d244b" exitCode=0 Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.371391 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s9ncx" event={"ID":"dbba7a9b-2c9b-4770-ac96-b89390457eab","Type":"ContainerDied","Data":"cf4922608027344a703c1ef9f815054a3bdba81a21392a04abf1e2d76d3d244b"} Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.371450 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s9ncx" event={"ID":"dbba7a9b-2c9b-4770-ac96-b89390457eab","Type":"ContainerStarted","Data":"184cf4856a6dd957f402dab4aa900b8d61954d494dad4611ca54a419a5cf33f1"} Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.375302 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" event={"ID":"b0a9d170-2b10-48f1-81ab-898a7df26740","Type":"ContainerStarted","Data":"075d1fc03e48621e182e44bb4396a0c0168b5e62216d32c6427804707a32e2df"} Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.378332 4923 generic.go:334] "Generic (PLEG): container finished" podID="7ce0bb4a-0841-4449-bc02-6c219731272e" containerID="1dde84cc50f06c1d02f0b31364a095423f98d2acad51e1a2cb5d8e255513a212" exitCode=0 Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.378453 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqdq8" event={"ID":"7ce0bb4a-0841-4449-bc02-6c219731272e","Type":"ContainerDied","Data":"1dde84cc50f06c1d02f0b31364a095423f98d2acad51e1a2cb5d8e255513a212"} Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.378530 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqdq8" event={"ID":"7ce0bb4a-0841-4449-bc02-6c219731272e","Type":"ContainerStarted","Data":"0f0311972eaa16edc13729f5c146404ff3a9210c4764b51789aa37602d767169"} Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.380389 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" event={"ID":"cfda8f36-49b9-4081-bbea-f311dd0a87a1","Type":"ContainerDied","Data":"5e3664da06955918680e8c88d95740647f4ed9785e064e41f9e6b45a84b01679"} Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.380418 4923 scope.go:117] "RemoveContainer" containerID="a23af8b5446190635a930538e80a26c355ca590fd3665b3e9da27144b5ad9381" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.380514 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.406182 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.406526 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-serving-cert\") pod \"route-controller-manager-56b676fd75-kcjj5\" (UID: \"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560\") " pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.406604 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9kzt\" (UniqueName: \"kubernetes.io/projected/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-kube-api-access-k9kzt\") pod \"route-controller-manager-56b676fd75-kcjj5\" (UID: \"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560\") " pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.406638 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-client-ca\") pod \"route-controller-manager-56b676fd75-kcjj5\" (UID: \"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560\") " pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.406681 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-config\") pod \"route-controller-manager-56b676fd75-kcjj5\" (UID: \"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560\") " pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.416216 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.423179 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-g25ck"] Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.428197 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-g25ck"] Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.469888 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-lc9xf" podStartSLOduration=11.469870614 podStartE2EDuration="11.469870614s" podCreationTimestamp="2026-03-11 13:24:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:34.468639004 +0000 UTC m=+122.150540718" watchObservedRunningTime="2026-03-11 13:24:34.469870614 +0000 UTC m=+122.151772328" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.486955 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-66d9d6d464-j9rb9"] Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.493601 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66d9d6d464-j9rb9"] Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.493705 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.496688 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.496976 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.497144 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.497355 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.498406 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.499104 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.503896 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.507933 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-serving-cert\") pod \"route-controller-manager-56b676fd75-kcjj5\" (UID: \"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560\") " pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.507997 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9kzt\" (UniqueName: \"kubernetes.io/projected/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-kube-api-access-k9kzt\") pod \"route-controller-manager-56b676fd75-kcjj5\" (UID: \"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560\") " pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.508021 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-client-ca\") pod \"route-controller-manager-56b676fd75-kcjj5\" (UID: \"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560\") " pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.508046 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-config\") pod \"route-controller-manager-56b676fd75-kcjj5\" (UID: \"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560\") " pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.509409 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-config\") pod \"route-controller-manager-56b676fd75-kcjj5\" (UID: \"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560\") " pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.512485 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-client-ca\") pod \"route-controller-manager-56b676fd75-kcjj5\" (UID: \"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560\") " pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.520979 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-serving-cert\") pod \"route-controller-manager-56b676fd75-kcjj5\" (UID: \"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560\") " pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.528411 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9kzt\" (UniqueName: \"kubernetes.io/projected/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-kube-api-access-k9kzt\") pod \"route-controller-manager-56b676fd75-kcjj5\" (UID: \"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560\") " pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.550445 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql"] Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.550510 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-qk4ql"] Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.555303 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jcncp"] Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.556407 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jcncp" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.559667 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.566251 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jcncp"] Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.598408 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.609669 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/87afdf29-c448-483d-9cd5-e6828425c59b-proxy-ca-bundles\") pod \"controller-manager-66d9d6d464-j9rb9\" (UID: \"87afdf29-c448-483d-9cd5-e6828425c59b\") " pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.609745 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94wrw\" (UniqueName: \"kubernetes.io/projected/87afdf29-c448-483d-9cd5-e6828425c59b-kube-api-access-94wrw\") pod \"controller-manager-66d9d6d464-j9rb9\" (UID: \"87afdf29-c448-483d-9cd5-e6828425c59b\") " pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.609796 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac5c84a5-f129-440d-8423-f8d86b00e3b9-catalog-content\") pod \"redhat-marketplace-jcncp\" (UID: \"ac5c84a5-f129-440d-8423-f8d86b00e3b9\") " pod="openshift-marketplace/redhat-marketplace-jcncp" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.609818 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87afdf29-c448-483d-9cd5-e6828425c59b-serving-cert\") pod \"controller-manager-66d9d6d464-j9rb9\" (UID: \"87afdf29-c448-483d-9cd5-e6828425c59b\") " pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.609853 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87afdf29-c448-483d-9cd5-e6828425c59b-config\") pod \"controller-manager-66d9d6d464-j9rb9\" (UID: \"87afdf29-c448-483d-9cd5-e6828425c59b\") " pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.609873 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/87afdf29-c448-483d-9cd5-e6828425c59b-client-ca\") pod \"controller-manager-66d9d6d464-j9rb9\" (UID: \"87afdf29-c448-483d-9cd5-e6828425c59b\") " pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.609897 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq82b\" (UniqueName: \"kubernetes.io/projected/ac5c84a5-f129-440d-8423-f8d86b00e3b9-kube-api-access-vq82b\") pod \"redhat-marketplace-jcncp\" (UID: \"ac5c84a5-f129-440d-8423-f8d86b00e3b9\") " pod="openshift-marketplace/redhat-marketplace-jcncp" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.609916 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac5c84a5-f129-440d-8423-f8d86b00e3b9-utilities\") pod \"redhat-marketplace-jcncp\" (UID: \"ac5c84a5-f129-440d-8423-f8d86b00e3b9\") " pod="openshift-marketplace/redhat-marketplace-jcncp" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.692628 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fwfgr"] Mar 11 13:24:34 crc kubenswrapper[4923]: W0311 13:24:34.701494 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb819b4e5_85bd_4a31_bc2e_a119da615675.slice/crio-5e152f856cc842a14bf2685d7d785ce4b1198d88a2087c5df0e7856a7ef962f6 WatchSource:0}: Error finding container 5e152f856cc842a14bf2685d7d785ce4b1198d88a2087c5df0e7856a7ef962f6: Status 404 returned error can't find the container with id 5e152f856cc842a14bf2685d7d785ce4b1198d88a2087c5df0e7856a7ef962f6 Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.711872 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/87afdf29-c448-483d-9cd5-e6828425c59b-proxy-ca-bundles\") pod \"controller-manager-66d9d6d464-j9rb9\" (UID: \"87afdf29-c448-483d-9cd5-e6828425c59b\") " pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.711952 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94wrw\" (UniqueName: \"kubernetes.io/projected/87afdf29-c448-483d-9cd5-e6828425c59b-kube-api-access-94wrw\") pod \"controller-manager-66d9d6d464-j9rb9\" (UID: \"87afdf29-c448-483d-9cd5-e6828425c59b\") " pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.712028 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac5c84a5-f129-440d-8423-f8d86b00e3b9-catalog-content\") pod \"redhat-marketplace-jcncp\" (UID: \"ac5c84a5-f129-440d-8423-f8d86b00e3b9\") " pod="openshift-marketplace/redhat-marketplace-jcncp" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.712071 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87afdf29-c448-483d-9cd5-e6828425c59b-serving-cert\") pod \"controller-manager-66d9d6d464-j9rb9\" (UID: \"87afdf29-c448-483d-9cd5-e6828425c59b\") " pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.712121 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87afdf29-c448-483d-9cd5-e6828425c59b-config\") pod \"controller-manager-66d9d6d464-j9rb9\" (UID: \"87afdf29-c448-483d-9cd5-e6828425c59b\") " pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.712155 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/87afdf29-c448-483d-9cd5-e6828425c59b-client-ca\") pod \"controller-manager-66d9d6d464-j9rb9\" (UID: \"87afdf29-c448-483d-9cd5-e6828425c59b\") " pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.712190 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq82b\" (UniqueName: \"kubernetes.io/projected/ac5c84a5-f129-440d-8423-f8d86b00e3b9-kube-api-access-vq82b\") pod \"redhat-marketplace-jcncp\" (UID: \"ac5c84a5-f129-440d-8423-f8d86b00e3b9\") " pod="openshift-marketplace/redhat-marketplace-jcncp" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.712236 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac5c84a5-f129-440d-8423-f8d86b00e3b9-utilities\") pod \"redhat-marketplace-jcncp\" (UID: \"ac5c84a5-f129-440d-8423-f8d86b00e3b9\") " pod="openshift-marketplace/redhat-marketplace-jcncp" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.712959 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac5c84a5-f129-440d-8423-f8d86b00e3b9-utilities\") pod \"redhat-marketplace-jcncp\" (UID: \"ac5c84a5-f129-440d-8423-f8d86b00e3b9\") " pod="openshift-marketplace/redhat-marketplace-jcncp" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.713703 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/87afdf29-c448-483d-9cd5-e6828425c59b-proxy-ca-bundles\") pod \"controller-manager-66d9d6d464-j9rb9\" (UID: \"87afdf29-c448-483d-9cd5-e6828425c59b\") " pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.713948 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/87afdf29-c448-483d-9cd5-e6828425c59b-client-ca\") pod \"controller-manager-66d9d6d464-j9rb9\" (UID: \"87afdf29-c448-483d-9cd5-e6828425c59b\") " pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.713953 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac5c84a5-f129-440d-8423-f8d86b00e3b9-catalog-content\") pod \"redhat-marketplace-jcncp\" (UID: \"ac5c84a5-f129-440d-8423-f8d86b00e3b9\") " pod="openshift-marketplace/redhat-marketplace-jcncp" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.714297 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87afdf29-c448-483d-9cd5-e6828425c59b-config\") pod \"controller-manager-66d9d6d464-j9rb9\" (UID: \"87afdf29-c448-483d-9cd5-e6828425c59b\") " pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.728093 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87afdf29-c448-483d-9cd5-e6828425c59b-serving-cert\") pod \"controller-manager-66d9d6d464-j9rb9\" (UID: \"87afdf29-c448-483d-9cd5-e6828425c59b\") " pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.728811 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq82b\" (UniqueName: \"kubernetes.io/projected/ac5c84a5-f129-440d-8423-f8d86b00e3b9-kube-api-access-vq82b\") pod \"redhat-marketplace-jcncp\" (UID: \"ac5c84a5-f129-440d-8423-f8d86b00e3b9\") " pod="openshift-marketplace/redhat-marketplace-jcncp" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.729023 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94wrw\" (UniqueName: \"kubernetes.io/projected/87afdf29-c448-483d-9cd5-e6828425c59b-kube-api-access-94wrw\") pod \"controller-manager-66d9d6d464-j9rb9\" (UID: \"87afdf29-c448-483d-9cd5-e6828425c59b\") " pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.811788 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5"] Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.817124 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" Mar 11 13:24:34 crc kubenswrapper[4923]: W0311 13:24:34.818757 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8acfb1c_a8e0_4f96_b1ba_7a3210b30560.slice/crio-02f3c4af7c39203780fa6b6ae3dda4f4c7aae099077eed500c8de75933f3367f WatchSource:0}: Error finding container 02f3c4af7c39203780fa6b6ae3dda4f4c7aae099077eed500c8de75933f3367f: Status 404 returned error can't find the container with id 02f3c4af7c39203780fa6b6ae3dda4f4c7aae099077eed500c8de75933f3367f Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.899848 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jcncp" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.948230 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mcgpl"] Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.950370 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mcgpl" Mar 11 13:24:34 crc kubenswrapper[4923]: I0311 13:24:34.956725 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mcgpl"] Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.017331 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnsfj\" (UniqueName: \"kubernetes.io/projected/9ab29276-3a67-42c9-8308-2ee71a3dd4dc-kube-api-access-hnsfj\") pod \"redhat-marketplace-mcgpl\" (UID: \"9ab29276-3a67-42c9-8308-2ee71a3dd4dc\") " pod="openshift-marketplace/redhat-marketplace-mcgpl" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.017422 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ab29276-3a67-42c9-8308-2ee71a3dd4dc-catalog-content\") pod \"redhat-marketplace-mcgpl\" (UID: \"9ab29276-3a67-42c9-8308-2ee71a3dd4dc\") " pod="openshift-marketplace/redhat-marketplace-mcgpl" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.017470 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ab29276-3a67-42c9-8308-2ee71a3dd4dc-utilities\") pod \"redhat-marketplace-mcgpl\" (UID: \"9ab29276-3a67-42c9-8308-2ee71a3dd4dc\") " pod="openshift-marketplace/redhat-marketplace-mcgpl" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.053053 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16d21cac-dc41-43fb-a391-8864085617ed" path="/var/lib/kubelet/pods/16d21cac-dc41-43fb-a391-8864085617ed/volumes" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.053801 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.054407 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfda8f36-49b9-4081-bbea-f311dd0a87a1" path="/var/lib/kubelet/pods/cfda8f36-49b9-4081-bbea-f311dd0a87a1/volumes" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.079174 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-66d9d6d464-j9rb9"] Mar 11 13:24:35 crc kubenswrapper[4923]: W0311 13:24:35.094777 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod87afdf29_c448_483d_9cd5_e6828425c59b.slice/crio-4aff2d39fd91159ce2e5e1a1f507339fb01f8928d79a34b62c096b9b3fafcaea WatchSource:0}: Error finding container 4aff2d39fd91159ce2e5e1a1f507339fb01f8928d79a34b62c096b9b3fafcaea: Status 404 returned error can't find the container with id 4aff2d39fd91159ce2e5e1a1f507339fb01f8928d79a34b62c096b9b3fafcaea Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.118749 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnsfj\" (UniqueName: \"kubernetes.io/projected/9ab29276-3a67-42c9-8308-2ee71a3dd4dc-kube-api-access-hnsfj\") pod \"redhat-marketplace-mcgpl\" (UID: \"9ab29276-3a67-42c9-8308-2ee71a3dd4dc\") " pod="openshift-marketplace/redhat-marketplace-mcgpl" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.118829 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ab29276-3a67-42c9-8308-2ee71a3dd4dc-catalog-content\") pod \"redhat-marketplace-mcgpl\" (UID: \"9ab29276-3a67-42c9-8308-2ee71a3dd4dc\") " pod="openshift-marketplace/redhat-marketplace-mcgpl" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.118864 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ab29276-3a67-42c9-8308-2ee71a3dd4dc-utilities\") pod \"redhat-marketplace-mcgpl\" (UID: \"9ab29276-3a67-42c9-8308-2ee71a3dd4dc\") " pod="openshift-marketplace/redhat-marketplace-mcgpl" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.119615 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ab29276-3a67-42c9-8308-2ee71a3dd4dc-utilities\") pod \"redhat-marketplace-mcgpl\" (UID: \"9ab29276-3a67-42c9-8308-2ee71a3dd4dc\") " pod="openshift-marketplace/redhat-marketplace-mcgpl" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.120171 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ab29276-3a67-42c9-8308-2ee71a3dd4dc-catalog-content\") pod \"redhat-marketplace-mcgpl\" (UID: \"9ab29276-3a67-42c9-8308-2ee71a3dd4dc\") " pod="openshift-marketplace/redhat-marketplace-mcgpl" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.151798 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.157661 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.161918 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.164437 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.166086 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.169027 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnsfj\" (UniqueName: \"kubernetes.io/projected/9ab29276-3a67-42c9-8308-2ee71a3dd4dc-kube-api-access-hnsfj\") pod \"redhat-marketplace-mcgpl\" (UID: \"9ab29276-3a67-42c9-8308-2ee71a3dd4dc\") " pod="openshift-marketplace/redhat-marketplace-mcgpl" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.206105 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.206892 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.210660 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.212146 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.212676 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.215068 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jcncp"] Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.220964 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/87bd3c68-b0fc-4008-ad02-920658fe0f01-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"87bd3c68-b0fc-4008-ad02-920658fe0f01\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.221075 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/87bd3c68-b0fc-4008-ad02-920658fe0f01-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"87bd3c68-b0fc-4008-ad02-920658fe0f01\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.221118 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b93878b7-e524-49b0-89af-dfe5de598a5e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b93878b7-e524-49b0-89af-dfe5de598a5e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.222931 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b93878b7-e524-49b0-89af-dfe5de598a5e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b93878b7-e524-49b0-89af-dfe5de598a5e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 11 13:24:35 crc kubenswrapper[4923]: W0311 13:24:35.242285 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac5c84a5_f129_440d_8423_f8d86b00e3b9.slice/crio-f66f05bc927a40ada2f03065ebbcf30bc630963cfe8ca94d4e96130f781fd07d WatchSource:0}: Error finding container f66f05bc927a40ada2f03065ebbcf30bc630963cfe8ca94d4e96130f781fd07d: Status 404 returned error can't find the container with id f66f05bc927a40ada2f03065ebbcf30bc630963cfe8ca94d4e96130f781fd07d Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.278124 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mcgpl" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.324241 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b93878b7-e524-49b0-89af-dfe5de598a5e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b93878b7-e524-49b0-89af-dfe5de598a5e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.324287 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/87bd3c68-b0fc-4008-ad02-920658fe0f01-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"87bd3c68-b0fc-4008-ad02-920658fe0f01\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.324324 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/87bd3c68-b0fc-4008-ad02-920658fe0f01-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"87bd3c68-b0fc-4008-ad02-920658fe0f01\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.324357 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b93878b7-e524-49b0-89af-dfe5de598a5e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b93878b7-e524-49b0-89af-dfe5de598a5e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.324462 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b93878b7-e524-49b0-89af-dfe5de598a5e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b93878b7-e524-49b0-89af-dfe5de598a5e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.324595 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/87bd3c68-b0fc-4008-ad02-920658fe0f01-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"87bd3c68-b0fc-4008-ad02-920658fe0f01\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.342420 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/87bd3c68-b0fc-4008-ad02-920658fe0f01-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"87bd3c68-b0fc-4008-ad02-920658fe0f01\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.344291 4923 patch_prober.go:28] interesting pod/router-default-5444994796-zd9ns container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 11 13:24:35 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Mar 11 13:24:35 crc kubenswrapper[4923]: [+]process-running ok Mar 11 13:24:35 crc kubenswrapper[4923]: healthz check failed Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.344324 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zd9ns" podUID="3f8730fb-a1a2-440e-8fb2-1bf9134054e2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.344294 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b93878b7-e524-49b0-89af-dfe5de598a5e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b93878b7-e524-49b0-89af-dfe5de598a5e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.413406 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" event={"ID":"87afdf29-c448-483d-9cd5-e6828425c59b","Type":"ContainerStarted","Data":"602f5459db5cfde4f8cb7949b672b34d021c30813b2ad553c98879beff34c495"} Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.413741 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" event={"ID":"87afdf29-c448-483d-9cd5-e6828425c59b","Type":"ContainerStarted","Data":"4aff2d39fd91159ce2e5e1a1f507339fb01f8928d79a34b62c096b9b3fafcaea"} Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.413759 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.417416 4923 patch_prober.go:28] interesting pod/controller-manager-66d9d6d464-j9rb9 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.49:8443/healthz\": dial tcp 10.217.0.49:8443: connect: connection refused" start-of-body= Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.417459 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" podUID="87afdf29-c448-483d-9cd5-e6828425c59b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.49:8443/healthz\": dial tcp 10.217.0.49:8443: connect: connection refused" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.429670 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" podStartSLOduration=2.42965419 podStartE2EDuration="2.42965419s" podCreationTimestamp="2026-03-11 13:24:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:35.427949727 +0000 UTC m=+123.109851441" watchObservedRunningTime="2026-03-11 13:24:35.42965419 +0000 UTC m=+123.111555904" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.440776 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jcncp" event={"ID":"ac5c84a5-f129-440d-8423-f8d86b00e3b9","Type":"ContainerStarted","Data":"57a50dd1eeb8b3ff16a98538327c923e7ed49047fea234d6cdbe1d8c02bd4903"} Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.440824 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jcncp" event={"ID":"ac5c84a5-f129-440d-8423-f8d86b00e3b9","Type":"ContainerStarted","Data":"f66f05bc927a40ada2f03065ebbcf30bc630963cfe8ca94d4e96130f781fd07d"} Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.445769 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" event={"ID":"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560","Type":"ContainerStarted","Data":"a22b4179139b9031d2712e5439138ae6419766d8b43c69e76d47e5f30c937b02"} Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.445798 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" event={"ID":"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560","Type":"ContainerStarted","Data":"02f3c4af7c39203780fa6b6ae3dda4f4c7aae099077eed500c8de75933f3367f"} Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.446375 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.452568 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" event={"ID":"b819b4e5-85bd-4a31-bc2e-a119da615675","Type":"ContainerStarted","Data":"924e8bb4b0d1749a95af14ab43eb2838fb04a0aeab2b3088e568931d3a87ec67"} Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.452600 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.452610 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" event={"ID":"b819b4e5-85bd-4a31-bc2e-a119da615675","Type":"ContainerStarted","Data":"5e152f856cc842a14bf2685d7d785ce4b1198d88a2087c5df0e7856a7ef962f6"} Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.484308 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" podStartSLOduration=70.484291155 podStartE2EDuration="1m10.484291155s" podCreationTimestamp="2026-03-11 13:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:35.480754838 +0000 UTC m=+123.162656572" watchObservedRunningTime="2026-03-11 13:24:35.484291155 +0000 UTC m=+123.166192869" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.494187 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" podStartSLOduration=1.494169361 podStartE2EDuration="1.494169361s" podCreationTimestamp="2026-03-11 13:24:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:35.493976136 +0000 UTC m=+123.175877850" watchObservedRunningTime="2026-03-11 13:24:35.494169361 +0000 UTC m=+123.176071075" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.551193 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rnjrh"] Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.552383 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rnjrh" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.554411 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.560195 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rnjrh"] Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.582378 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.614834 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.629278 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/094d43b0-8aa7-4acc-9d97-9c00874ee1a8-catalog-content\") pod \"redhat-operators-rnjrh\" (UID: \"094d43b0-8aa7-4acc-9d97-9c00874ee1a8\") " pod="openshift-marketplace/redhat-operators-rnjrh" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.629405 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x55th\" (UniqueName: \"kubernetes.io/projected/094d43b0-8aa7-4acc-9d97-9c00874ee1a8-kube-api-access-x55th\") pod \"redhat-operators-rnjrh\" (UID: \"094d43b0-8aa7-4acc-9d97-9c00874ee1a8\") " pod="openshift-marketplace/redhat-operators-rnjrh" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.629463 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/094d43b0-8aa7-4acc-9d97-9c00874ee1a8-utilities\") pod \"redhat-operators-rnjrh\" (UID: \"094d43b0-8aa7-4acc-9d97-9c00874ee1a8\") " pod="openshift-marketplace/redhat-operators-rnjrh" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.643506 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.644232 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.660371 4923 patch_prober.go:28] interesting pod/console-f9d7485db-mgrcg container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.660414 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-mgrcg" podUID="13238a87-3b9a-471a-8337-7e7f15018221" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.660560 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mcgpl"] Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.662498 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.663428 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.690925 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.732627 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/094d43b0-8aa7-4acc-9d97-9c00874ee1a8-catalog-content\") pod \"redhat-operators-rnjrh\" (UID: \"094d43b0-8aa7-4acc-9d97-9c00874ee1a8\") " pod="openshift-marketplace/redhat-operators-rnjrh" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.732972 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x55th\" (UniqueName: \"kubernetes.io/projected/094d43b0-8aa7-4acc-9d97-9c00874ee1a8-kube-api-access-x55th\") pod \"redhat-operators-rnjrh\" (UID: \"094d43b0-8aa7-4acc-9d97-9c00874ee1a8\") " pod="openshift-marketplace/redhat-operators-rnjrh" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.733027 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/094d43b0-8aa7-4acc-9d97-9c00874ee1a8-utilities\") pod \"redhat-operators-rnjrh\" (UID: \"094d43b0-8aa7-4acc-9d97-9c00874ee1a8\") " pod="openshift-marketplace/redhat-operators-rnjrh" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.736078 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/094d43b0-8aa7-4acc-9d97-9c00874ee1a8-catalog-content\") pod \"redhat-operators-rnjrh\" (UID: \"094d43b0-8aa7-4acc-9d97-9c00874ee1a8\") " pod="openshift-marketplace/redhat-operators-rnjrh" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.736566 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/094d43b0-8aa7-4acc-9d97-9c00874ee1a8-utilities\") pod \"redhat-operators-rnjrh\" (UID: \"094d43b0-8aa7-4acc-9d97-9c00874ee1a8\") " pod="openshift-marketplace/redhat-operators-rnjrh" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.757807 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x55th\" (UniqueName: \"kubernetes.io/projected/094d43b0-8aa7-4acc-9d97-9c00874ee1a8-kube-api-access-x55th\") pod \"redhat-operators-rnjrh\" (UID: \"094d43b0-8aa7-4acc-9d97-9c00874ee1a8\") " pod="openshift-marketplace/redhat-operators-rnjrh" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.818697 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.868833 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rnjrh" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.913781 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.914067 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.924258 4923 patch_prober.go:28] interesting pod/apiserver-76f77b778f-k6dxb container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Mar 11 13:24:35 crc kubenswrapper[4923]: [+]log ok Mar 11 13:24:35 crc kubenswrapper[4923]: [+]etcd ok Mar 11 13:24:35 crc kubenswrapper[4923]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 11 13:24:35 crc kubenswrapper[4923]: [+]poststarthook/generic-apiserver-start-informers ok Mar 11 13:24:35 crc kubenswrapper[4923]: [+]poststarthook/max-in-flight-filter ok Mar 11 13:24:35 crc kubenswrapper[4923]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 11 13:24:35 crc kubenswrapper[4923]: [+]poststarthook/image.openshift.io-apiserver-caches ok Mar 11 13:24:35 crc kubenswrapper[4923]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Mar 11 13:24:35 crc kubenswrapper[4923]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Mar 11 13:24:35 crc kubenswrapper[4923]: [+]poststarthook/project.openshift.io-projectcache ok Mar 11 13:24:35 crc kubenswrapper[4923]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Mar 11 13:24:35 crc kubenswrapper[4923]: [+]poststarthook/openshift.io-startinformers ok Mar 11 13:24:35 crc kubenswrapper[4923]: [+]poststarthook/openshift.io-restmapperupdater ok Mar 11 13:24:35 crc kubenswrapper[4923]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Mar 11 13:24:35 crc kubenswrapper[4923]: livez check failed Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.924323 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" podUID="1515573d-f360-41f2-854a-4d0fb437af7d" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.951181 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-97n98"] Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.952246 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-97n98" Mar 11 13:24:35 crc kubenswrapper[4923]: I0311 13:24:35.990893 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.021461 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-97n98"] Mar 11 13:24:36 crc kubenswrapper[4923]: W0311 13:24:36.110485 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod87bd3c68_b0fc_4008_ad02_920658fe0f01.slice/crio-94f55149b40ed1a5975a5b2543908ffd66a11b54df13a57512d224cdcee3437c WatchSource:0}: Error finding container 94f55149b40ed1a5975a5b2543908ffd66a11b54df13a57512d224cdcee3437c: Status 404 returned error can't find the container with id 94f55149b40ed1a5975a5b2543908ffd66a11b54df13a57512d224cdcee3437c Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.138241 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s29bn\" (UniqueName: \"kubernetes.io/projected/08394455-c98e-4e73-b798-3a9c2ff5e708-kube-api-access-s29bn\") pod \"redhat-operators-97n98\" (UID: \"08394455-c98e-4e73-b798-3a9c2ff5e708\") " pod="openshift-marketplace/redhat-operators-97n98" Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.138317 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08394455-c98e-4e73-b798-3a9c2ff5e708-utilities\") pod \"redhat-operators-97n98\" (UID: \"08394455-c98e-4e73-b798-3a9c2ff5e708\") " pod="openshift-marketplace/redhat-operators-97n98" Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.138385 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08394455-c98e-4e73-b798-3a9c2ff5e708-catalog-content\") pod \"redhat-operators-97n98\" (UID: \"08394455-c98e-4e73-b798-3a9c2ff5e708\") " pod="openshift-marketplace/redhat-operators-97n98" Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.174018 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.246129 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s29bn\" (UniqueName: \"kubernetes.io/projected/08394455-c98e-4e73-b798-3a9c2ff5e708-kube-api-access-s29bn\") pod \"redhat-operators-97n98\" (UID: \"08394455-c98e-4e73-b798-3a9c2ff5e708\") " pod="openshift-marketplace/redhat-operators-97n98" Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.246505 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08394455-c98e-4e73-b798-3a9c2ff5e708-utilities\") pod \"redhat-operators-97n98\" (UID: \"08394455-c98e-4e73-b798-3a9c2ff5e708\") " pod="openshift-marketplace/redhat-operators-97n98" Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.246559 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08394455-c98e-4e73-b798-3a9c2ff5e708-catalog-content\") pod \"redhat-operators-97n98\" (UID: \"08394455-c98e-4e73-b798-3a9c2ff5e708\") " pod="openshift-marketplace/redhat-operators-97n98" Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.247250 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08394455-c98e-4e73-b798-3a9c2ff5e708-utilities\") pod \"redhat-operators-97n98\" (UID: \"08394455-c98e-4e73-b798-3a9c2ff5e708\") " pod="openshift-marketplace/redhat-operators-97n98" Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.247474 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08394455-c98e-4e73-b798-3a9c2ff5e708-catalog-content\") pod \"redhat-operators-97n98\" (UID: \"08394455-c98e-4e73-b798-3a9c2ff5e708\") " pod="openshift-marketplace/redhat-operators-97n98" Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.260405 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rnjrh"] Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.275862 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s29bn\" (UniqueName: \"kubernetes.io/projected/08394455-c98e-4e73-b798-3a9c2ff5e708-kube-api-access-s29bn\") pod \"redhat-operators-97n98\" (UID: \"08394455-c98e-4e73-b798-3a9c2ff5e708\") " pod="openshift-marketplace/redhat-operators-97n98" Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.339722 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-zd9ns" Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.344026 4923 patch_prober.go:28] interesting pod/router-default-5444994796-zd9ns container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 11 13:24:36 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Mar 11 13:24:36 crc kubenswrapper[4923]: [+]process-running ok Mar 11 13:24:36 crc kubenswrapper[4923]: healthz check failed Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.344067 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zd9ns" podUID="3f8730fb-a1a2-440e-8fb2-1bf9134054e2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.453096 4923 ???:1] "http: TLS handshake error from 192.168.126.11:39034: no serving certificate available for the kubelet" Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.484427 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"87bd3c68-b0fc-4008-ad02-920658fe0f01","Type":"ContainerStarted","Data":"94f55149b40ed1a5975a5b2543908ffd66a11b54df13a57512d224cdcee3437c"} Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.493702 4923 generic.go:334] "Generic (PLEG): container finished" podID="ac5c84a5-f129-440d-8423-f8d86b00e3b9" containerID="57a50dd1eeb8b3ff16a98538327c923e7ed49047fea234d6cdbe1d8c02bd4903" exitCode=0 Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.493778 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jcncp" event={"ID":"ac5c84a5-f129-440d-8423-f8d86b00e3b9","Type":"ContainerDied","Data":"57a50dd1eeb8b3ff16a98538327c923e7ed49047fea234d6cdbe1d8c02bd4903"} Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.496309 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnjrh" event={"ID":"094d43b0-8aa7-4acc-9d97-9c00874ee1a8","Type":"ContainerStarted","Data":"58bf1236c788c6461942b987737349e09a6c530f01f072baa7e67713d554af3f"} Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.497312 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b93878b7-e524-49b0-89af-dfe5de598a5e","Type":"ContainerStarted","Data":"89ab1450ceb8e22291251007d135b070e3ede0bef0be652d29f335aea5ca8ba2"} Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.501052 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mcgpl" event={"ID":"9ab29276-3a67-42c9-8308-2ee71a3dd4dc","Type":"ContainerStarted","Data":"c8462f8718db09f8a61636a049de2e143ca3bf015bf2d3f505f39ac804f9d295"} Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.501099 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mcgpl" event={"ID":"9ab29276-3a67-42c9-8308-2ee71a3dd4dc","Type":"ContainerStarted","Data":"4267893edf7067807fb0c8157225ea1ab1879a1bc50e3c7f871d804888d2565b"} Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.505786 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.507150 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-rqlvv" Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.567749 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-97n98" Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.648583 4923 ???:1] "http: TLS handshake error from 192.168.126.11:39040: no serving certificate available for the kubelet" Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.747044 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" Mar 11 13:24:36 crc kubenswrapper[4923]: E0311 13:24:36.760446 4923 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d348f00e8a6e4b07e7a23195b2d70fdf885db23f4ce1b740a3fcb77ccd7756ec" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 11 13:24:36 crc kubenswrapper[4923]: E0311 13:24:36.764415 4923 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d348f00e8a6e4b07e7a23195b2d70fdf885db23f4ce1b740a3fcb77ccd7756ec" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 11 13:24:36 crc kubenswrapper[4923]: E0311 13:24:36.788981 4923 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d348f00e8a6e4b07e7a23195b2d70fdf885db23f4ce1b740a3fcb77ccd7756ec" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 11 13:24:36 crc kubenswrapper[4923]: E0311 13:24:36.789048 4923 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" podUID="a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04" containerName="kube-multus-additional-cni-plugins" Mar 11 13:24:36 crc kubenswrapper[4923]: I0311 13:24:36.930401 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-97n98"] Mar 11 13:24:37 crc kubenswrapper[4923]: I0311 13:24:37.343687 4923 patch_prober.go:28] interesting pod/router-default-5444994796-zd9ns container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 11 13:24:37 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Mar 11 13:24:37 crc kubenswrapper[4923]: [+]process-running ok Mar 11 13:24:37 crc kubenswrapper[4923]: healthz check failed Mar 11 13:24:37 crc kubenswrapper[4923]: I0311 13:24:37.343761 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zd9ns" podUID="3f8730fb-a1a2-440e-8fb2-1bf9134054e2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 11 13:24:37 crc kubenswrapper[4923]: I0311 13:24:37.511446 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-97n98" event={"ID":"08394455-c98e-4e73-b798-3a9c2ff5e708","Type":"ContainerStarted","Data":"298cf8caabfb130792a48b857e4a438ce031ac0223096788367622e427775fbf"} Mar 11 13:24:37 crc kubenswrapper[4923]: I0311 13:24:37.513690 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"87bd3c68-b0fc-4008-ad02-920658fe0f01","Type":"ContainerStarted","Data":"63c9ef14fe2a0d5a7479f9d66f68df188a2b8e515d36815addf9456e44396788"} Mar 11 13:24:37 crc kubenswrapper[4923]: I0311 13:24:37.516511 4923 generic.go:334] "Generic (PLEG): container finished" podID="9ab29276-3a67-42c9-8308-2ee71a3dd4dc" containerID="c8462f8718db09f8a61636a049de2e143ca3bf015bf2d3f505f39ac804f9d295" exitCode=0 Mar 11 13:24:37 crc kubenswrapper[4923]: I0311 13:24:37.516585 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mcgpl" event={"ID":"9ab29276-3a67-42c9-8308-2ee71a3dd4dc","Type":"ContainerDied","Data":"c8462f8718db09f8a61636a049de2e143ca3bf015bf2d3f505f39ac804f9d295"} Mar 11 13:24:37 crc kubenswrapper[4923]: I0311 13:24:37.521460 4923 generic.go:334] "Generic (PLEG): container finished" podID="094d43b0-8aa7-4acc-9d97-9c00874ee1a8" containerID="d78a8ffda1ef27c83eadbf30b6f491cfb8d2682a01577dfabdd6a0165e9db445" exitCode=0 Mar 11 13:24:37 crc kubenswrapper[4923]: I0311 13:24:37.521602 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnjrh" event={"ID":"094d43b0-8aa7-4acc-9d97-9c00874ee1a8","Type":"ContainerDied","Data":"d78a8ffda1ef27c83eadbf30b6f491cfb8d2682a01577dfabdd6a0165e9db445"} Mar 11 13:24:38 crc kubenswrapper[4923]: I0311 13:24:38.033489 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 11 13:24:38 crc kubenswrapper[4923]: I0311 13:24:38.333708 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-7lx27" Mar 11 13:24:38 crc kubenswrapper[4923]: I0311 13:24:38.342096 4923 patch_prober.go:28] interesting pod/router-default-5444994796-zd9ns container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 11 13:24:38 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Mar 11 13:24:38 crc kubenswrapper[4923]: [+]process-running ok Mar 11 13:24:38 crc kubenswrapper[4923]: healthz check failed Mar 11 13:24:38 crc kubenswrapper[4923]: I0311 13:24:38.342140 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zd9ns" podUID="3f8730fb-a1a2-440e-8fb2-1bf9134054e2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 11 13:24:38 crc kubenswrapper[4923]: I0311 13:24:38.350620 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=0.35060264 podStartE2EDuration="350.60264ms" podCreationTimestamp="2026-03-11 13:24:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:24:38.349899172 +0000 UTC m=+126.031800886" watchObservedRunningTime="2026-03-11 13:24:38.35060264 +0000 UTC m=+126.032504354" Mar 11 13:24:38 crc kubenswrapper[4923]: I0311 13:24:38.533512 4923 generic.go:334] "Generic (PLEG): container finished" podID="b93878b7-e524-49b0-89af-dfe5de598a5e" containerID="4d7e2bb27d9f1bb159d8e92567771a419e7c0b69261b6fe4210e71842c19b158" exitCode=0 Mar 11 13:24:38 crc kubenswrapper[4923]: I0311 13:24:38.533586 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b93878b7-e524-49b0-89af-dfe5de598a5e","Type":"ContainerDied","Data":"4d7e2bb27d9f1bb159d8e92567771a419e7c0b69261b6fe4210e71842c19b158"} Mar 11 13:24:38 crc kubenswrapper[4923]: I0311 13:24:38.536064 4923 generic.go:334] "Generic (PLEG): container finished" podID="08394455-c98e-4e73-b798-3a9c2ff5e708" containerID="c14e06ae906be3c2a21149ec1115cd7fd6b13cb240a32bffd813fae42269b678" exitCode=0 Mar 11 13:24:38 crc kubenswrapper[4923]: I0311 13:24:38.536166 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-97n98" event={"ID":"08394455-c98e-4e73-b798-3a9c2ff5e708","Type":"ContainerDied","Data":"c14e06ae906be3c2a21149ec1115cd7fd6b13cb240a32bffd813fae42269b678"} Mar 11 13:24:38 crc kubenswrapper[4923]: I0311 13:24:38.538053 4923 generic.go:334] "Generic (PLEG): container finished" podID="87bd3c68-b0fc-4008-ad02-920658fe0f01" containerID="63c9ef14fe2a0d5a7479f9d66f68df188a2b8e515d36815addf9456e44396788" exitCode=0 Mar 11 13:24:38 crc kubenswrapper[4923]: I0311 13:24:38.538162 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"87bd3c68-b0fc-4008-ad02-920658fe0f01","Type":"ContainerDied","Data":"63c9ef14fe2a0d5a7479f9d66f68df188a2b8e515d36815addf9456e44396788"} Mar 11 13:24:39 crc kubenswrapper[4923]: I0311 13:24:39.342314 4923 patch_prober.go:28] interesting pod/router-default-5444994796-zd9ns container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 11 13:24:39 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Mar 11 13:24:39 crc kubenswrapper[4923]: [+]process-running ok Mar 11 13:24:39 crc kubenswrapper[4923]: healthz check failed Mar 11 13:24:39 crc kubenswrapper[4923]: I0311 13:24:39.342379 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zd9ns" podUID="3f8730fb-a1a2-440e-8fb2-1bf9134054e2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 11 13:24:40 crc kubenswrapper[4923]: I0311 13:24:40.342177 4923 patch_prober.go:28] interesting pod/router-default-5444994796-zd9ns container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 11 13:24:40 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Mar 11 13:24:40 crc kubenswrapper[4923]: [+]process-running ok Mar 11 13:24:40 crc kubenswrapper[4923]: healthz check failed Mar 11 13:24:40 crc kubenswrapper[4923]: I0311 13:24:40.342457 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zd9ns" podUID="3f8730fb-a1a2-440e-8fb2-1bf9134054e2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 11 13:24:40 crc kubenswrapper[4923]: I0311 13:24:40.917062 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:40 crc kubenswrapper[4923]: I0311 13:24:40.922488 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-k6dxb" Mar 11 13:24:41 crc kubenswrapper[4923]: I0311 13:24:41.342611 4923 patch_prober.go:28] interesting pod/router-default-5444994796-zd9ns container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 11 13:24:41 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Mar 11 13:24:41 crc kubenswrapper[4923]: [+]process-running ok Mar 11 13:24:41 crc kubenswrapper[4923]: healthz check failed Mar 11 13:24:41 crc kubenswrapper[4923]: I0311 13:24:41.342830 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zd9ns" podUID="3f8730fb-a1a2-440e-8fb2-1bf9134054e2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 11 13:24:41 crc kubenswrapper[4923]: I0311 13:24:41.594962 4923 ???:1] "http: TLS handshake error from 192.168.126.11:53764: no serving certificate available for the kubelet" Mar 11 13:24:42 crc kubenswrapper[4923]: I0311 13:24:42.342801 4923 patch_prober.go:28] interesting pod/router-default-5444994796-zd9ns container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 11 13:24:42 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Mar 11 13:24:42 crc kubenswrapper[4923]: [+]process-running ok Mar 11 13:24:42 crc kubenswrapper[4923]: healthz check failed Mar 11 13:24:42 crc kubenswrapper[4923]: I0311 13:24:42.342851 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zd9ns" podUID="3f8730fb-a1a2-440e-8fb2-1bf9134054e2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 11 13:24:42 crc kubenswrapper[4923]: I0311 13:24:42.377354 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:24:43 crc kubenswrapper[4923]: I0311 13:24:43.342163 4923 patch_prober.go:28] interesting pod/router-default-5444994796-zd9ns container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 11 13:24:43 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Mar 11 13:24:43 crc kubenswrapper[4923]: [+]process-running ok Mar 11 13:24:43 crc kubenswrapper[4923]: healthz check failed Mar 11 13:24:43 crc kubenswrapper[4923]: I0311 13:24:43.342235 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zd9ns" podUID="3f8730fb-a1a2-440e-8fb2-1bf9134054e2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 11 13:24:43 crc kubenswrapper[4923]: I0311 13:24:43.702557 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:24:44 crc kubenswrapper[4923]: I0311 13:24:44.208917 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-b5jmc" Mar 11 13:24:44 crc kubenswrapper[4923]: I0311 13:24:44.341737 4923 patch_prober.go:28] interesting pod/router-default-5444994796-zd9ns container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 11 13:24:44 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Mar 11 13:24:44 crc kubenswrapper[4923]: [+]process-running ok Mar 11 13:24:44 crc kubenswrapper[4923]: healthz check failed Mar 11 13:24:44 crc kubenswrapper[4923]: I0311 13:24:44.341811 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zd9ns" podUID="3f8730fb-a1a2-440e-8fb2-1bf9134054e2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 11 13:24:45 crc kubenswrapper[4923]: I0311 13:24:45.348368 4923 patch_prober.go:28] interesting pod/router-default-5444994796-zd9ns container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 11 13:24:45 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Mar 11 13:24:45 crc kubenswrapper[4923]: [+]process-running ok Mar 11 13:24:45 crc kubenswrapper[4923]: healthz check failed Mar 11 13:24:45 crc kubenswrapper[4923]: I0311 13:24:45.348460 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zd9ns" podUID="3f8730fb-a1a2-440e-8fb2-1bf9134054e2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 11 13:24:45 crc kubenswrapper[4923]: I0311 13:24:45.644581 4923 patch_prober.go:28] interesting pod/console-f9d7485db-mgrcg container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Mar 11 13:24:45 crc kubenswrapper[4923]: I0311 13:24:45.644663 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-mgrcg" podUID="13238a87-3b9a-471a-8337-7e7f15018221" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Mar 11 13:24:45 crc kubenswrapper[4923]: I0311 13:24:45.964249 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 11 13:24:46 crc kubenswrapper[4923]: I0311 13:24:46.120319 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/87bd3c68-b0fc-4008-ad02-920658fe0f01-kube-api-access\") pod \"87bd3c68-b0fc-4008-ad02-920658fe0f01\" (UID: \"87bd3c68-b0fc-4008-ad02-920658fe0f01\") " Mar 11 13:24:46 crc kubenswrapper[4923]: I0311 13:24:46.120400 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/87bd3c68-b0fc-4008-ad02-920658fe0f01-kubelet-dir\") pod \"87bd3c68-b0fc-4008-ad02-920658fe0f01\" (UID: \"87bd3c68-b0fc-4008-ad02-920658fe0f01\") " Mar 11 13:24:46 crc kubenswrapper[4923]: I0311 13:24:46.120648 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87bd3c68-b0fc-4008-ad02-920658fe0f01-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "87bd3c68-b0fc-4008-ad02-920658fe0f01" (UID: "87bd3c68-b0fc-4008-ad02-920658fe0f01"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:24:46 crc kubenswrapper[4923]: I0311 13:24:46.128907 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87bd3c68-b0fc-4008-ad02-920658fe0f01-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "87bd3c68-b0fc-4008-ad02-920658fe0f01" (UID: "87bd3c68-b0fc-4008-ad02-920658fe0f01"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:46 crc kubenswrapper[4923]: I0311 13:24:46.222602 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/87bd3c68-b0fc-4008-ad02-920658fe0f01-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:46 crc kubenswrapper[4923]: I0311 13:24:46.222975 4923 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/87bd3c68-b0fc-4008-ad02-920658fe0f01-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:46 crc kubenswrapper[4923]: I0311 13:24:46.343089 4923 patch_prober.go:28] interesting pod/router-default-5444994796-zd9ns container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 11 13:24:46 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Mar 11 13:24:46 crc kubenswrapper[4923]: [+]process-running ok Mar 11 13:24:46 crc kubenswrapper[4923]: healthz check failed Mar 11 13:24:46 crc kubenswrapper[4923]: I0311 13:24:46.343159 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zd9ns" podUID="3f8730fb-a1a2-440e-8fb2-1bf9134054e2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 11 13:24:46 crc kubenswrapper[4923]: I0311 13:24:46.617394 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"87bd3c68-b0fc-4008-ad02-920658fe0f01","Type":"ContainerDied","Data":"94f55149b40ed1a5975a5b2543908ffd66a11b54df13a57512d224cdcee3437c"} Mar 11 13:24:46 crc kubenswrapper[4923]: I0311 13:24:46.617433 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="94f55149b40ed1a5975a5b2543908ffd66a11b54df13a57512d224cdcee3437c" Mar 11 13:24:46 crc kubenswrapper[4923]: I0311 13:24:46.617479 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 11 13:24:46 crc kubenswrapper[4923]: E0311 13:24:46.760187 4923 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d348f00e8a6e4b07e7a23195b2d70fdf885db23f4ce1b740a3fcb77ccd7756ec" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 11 13:24:46 crc kubenswrapper[4923]: E0311 13:24:46.761956 4923 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d348f00e8a6e4b07e7a23195b2d70fdf885db23f4ce1b740a3fcb77ccd7756ec" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 11 13:24:46 crc kubenswrapper[4923]: E0311 13:24:46.763420 4923 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d348f00e8a6e4b07e7a23195b2d70fdf885db23f4ce1b740a3fcb77ccd7756ec" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 11 13:24:46 crc kubenswrapper[4923]: E0311 13:24:46.763460 4923 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" podUID="a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04" containerName="kube-multus-additional-cni-plugins" Mar 11 13:24:47 crc kubenswrapper[4923]: I0311 13:24:47.324760 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 11 13:24:47 crc kubenswrapper[4923]: I0311 13:24:47.348966 4923 patch_prober.go:28] interesting pod/router-default-5444994796-zd9ns container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 11 13:24:47 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Mar 11 13:24:47 crc kubenswrapper[4923]: [+]process-running ok Mar 11 13:24:47 crc kubenswrapper[4923]: healthz check failed Mar 11 13:24:47 crc kubenswrapper[4923]: I0311 13:24:47.349033 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-zd9ns" podUID="3f8730fb-a1a2-440e-8fb2-1bf9134054e2" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 11 13:24:47 crc kubenswrapper[4923]: I0311 13:24:47.438447 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b93878b7-e524-49b0-89af-dfe5de598a5e-kubelet-dir\") pod \"b93878b7-e524-49b0-89af-dfe5de598a5e\" (UID: \"b93878b7-e524-49b0-89af-dfe5de598a5e\") " Mar 11 13:24:47 crc kubenswrapper[4923]: I0311 13:24:47.438550 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b93878b7-e524-49b0-89af-dfe5de598a5e-kube-api-access\") pod \"b93878b7-e524-49b0-89af-dfe5de598a5e\" (UID: \"b93878b7-e524-49b0-89af-dfe5de598a5e\") " Mar 11 13:24:47 crc kubenswrapper[4923]: I0311 13:24:47.438589 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b93878b7-e524-49b0-89af-dfe5de598a5e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b93878b7-e524-49b0-89af-dfe5de598a5e" (UID: "b93878b7-e524-49b0-89af-dfe5de598a5e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:24:47 crc kubenswrapper[4923]: I0311 13:24:47.438747 4923 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b93878b7-e524-49b0-89af-dfe5de598a5e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:47 crc kubenswrapper[4923]: I0311 13:24:47.456957 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b93878b7-e524-49b0-89af-dfe5de598a5e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b93878b7-e524-49b0-89af-dfe5de598a5e" (UID: "b93878b7-e524-49b0-89af-dfe5de598a5e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:24:47 crc kubenswrapper[4923]: I0311 13:24:47.539463 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b93878b7-e524-49b0-89af-dfe5de598a5e-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 11 13:24:47 crc kubenswrapper[4923]: I0311 13:24:47.625618 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"b93878b7-e524-49b0-89af-dfe5de598a5e","Type":"ContainerDied","Data":"89ab1450ceb8e22291251007d135b070e3ede0bef0be652d29f335aea5ca8ba2"} Mar 11 13:24:47 crc kubenswrapper[4923]: I0311 13:24:47.625657 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="89ab1450ceb8e22291251007d135b070e3ede0bef0be652d29f335aea5ca8ba2" Mar 11 13:24:47 crc kubenswrapper[4923]: I0311 13:24:47.625879 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 11 13:24:48 crc kubenswrapper[4923]: I0311 13:24:48.342881 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-zd9ns" Mar 11 13:24:48 crc kubenswrapper[4923]: I0311 13:24:48.345877 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-zd9ns" Mar 11 13:24:52 crc kubenswrapper[4923]: I0311 13:24:52.083717 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-66d9d6d464-j9rb9"] Mar 11 13:24:52 crc kubenswrapper[4923]: I0311 13:24:52.084647 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" podUID="87afdf29-c448-483d-9cd5-e6828425c59b" containerName="controller-manager" containerID="cri-o://602f5459db5cfde4f8cb7949b672b34d021c30813b2ad553c98879beff34c495" gracePeriod=30 Mar 11 13:24:52 crc kubenswrapper[4923]: I0311 13:24:52.105617 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5"] Mar 11 13:24:52 crc kubenswrapper[4923]: I0311 13:24:52.105905 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" podUID="d8acfb1c-a8e0-4f96-b1ba-7a3210b30560" containerName="route-controller-manager" containerID="cri-o://a22b4179139b9031d2712e5439138ae6419766d8b43c69e76d47e5f30c937b02" gracePeriod=30 Mar 11 13:24:53 crc kubenswrapper[4923]: I0311 13:24:53.661113 4923 generic.go:334] "Generic (PLEG): container finished" podID="d8acfb1c-a8e0-4f96-b1ba-7a3210b30560" containerID="a22b4179139b9031d2712e5439138ae6419766d8b43c69e76d47e5f30c937b02" exitCode=0 Mar 11 13:24:53 crc kubenswrapper[4923]: I0311 13:24:53.661219 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" event={"ID":"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560","Type":"ContainerDied","Data":"a22b4179139b9031d2712e5439138ae6419766d8b43c69e76d47e5f30c937b02"} Mar 11 13:24:53 crc kubenswrapper[4923]: I0311 13:24:53.663197 4923 generic.go:334] "Generic (PLEG): container finished" podID="87afdf29-c448-483d-9cd5-e6828425c59b" containerID="602f5459db5cfde4f8cb7949b672b34d021c30813b2ad553c98879beff34c495" exitCode=0 Mar 11 13:24:53 crc kubenswrapper[4923]: I0311 13:24:53.663249 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" event={"ID":"87afdf29-c448-483d-9cd5-e6828425c59b","Type":"ContainerDied","Data":"602f5459db5cfde4f8cb7949b672b34d021c30813b2ad553c98879beff34c495"} Mar 11 13:24:54 crc kubenswrapper[4923]: I0311 13:24:54.422190 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:24:54 crc kubenswrapper[4923]: I0311 13:24:54.599170 4923 patch_prober.go:28] interesting pod/route-controller-manager-56b676fd75-kcjj5 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.48:8443/healthz\": dial tcp 10.217.0.48:8443: connect: connection refused" start-of-body= Mar 11 13:24:54 crc kubenswrapper[4923]: I0311 13:24:54.599213 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" podUID="d8acfb1c-a8e0-4f96-b1ba-7a3210b30560" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.48:8443/healthz\": dial tcp 10.217.0.48:8443: connect: connection refused" Mar 11 13:24:55 crc kubenswrapper[4923]: I0311 13:24:55.652864 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:55 crc kubenswrapper[4923]: I0311 13:24:55.659925 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-mgrcg" Mar 11 13:24:55 crc kubenswrapper[4923]: I0311 13:24:55.819097 4923 patch_prober.go:28] interesting pod/controller-manager-66d9d6d464-j9rb9 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.49:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 11 13:24:55 crc kubenswrapper[4923]: I0311 13:24:55.819479 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" podUID="87afdf29-c448-483d-9cd5-e6828425c59b" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.49:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 11 13:24:56 crc kubenswrapper[4923]: E0311 13:24:56.761143 4923 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d348f00e8a6e4b07e7a23195b2d70fdf885db23f4ce1b740a3fcb77ccd7756ec" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 11 13:24:56 crc kubenswrapper[4923]: E0311 13:24:56.762844 4923 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d348f00e8a6e4b07e7a23195b2d70fdf885db23f4ce1b740a3fcb77ccd7756ec" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 11 13:24:56 crc kubenswrapper[4923]: E0311 13:24:56.764369 4923 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d348f00e8a6e4b07e7a23195b2d70fdf885db23f4ce1b740a3fcb77ccd7756ec" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 11 13:24:56 crc kubenswrapper[4923]: E0311 13:24:56.764406 4923 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" podUID="a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04" containerName="kube-multus-additional-cni-plugins" Mar 11 13:24:57 crc kubenswrapper[4923]: I0311 13:24:57.052997 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 11 13:25:01 crc kubenswrapper[4923]: E0311 13:25:01.761727 4923 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 11 13:25:01 crc kubenswrapper[4923]: E0311 13:25:01.762423 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d48t2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-8f4cb_openshift-marketplace(52f59385-23ad-4163-a766-b9d21e57ad6c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 11 13:25:01 crc kubenswrapper[4923]: E0311 13:25:01.763551 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-8f4cb" podUID="52f59385-23ad-4163-a766-b9d21e57ad6c" Mar 11 13:25:01 crc kubenswrapper[4923]: E0311 13:25:01.766910 4923 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 11 13:25:01 crc kubenswrapper[4923]: E0311 13:25:01.767398 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-drk5r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-rdvss_openshift-marketplace(8bdd5510-3bb1-42f0-8646-c45f11bdeafe): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 11 13:25:01 crc kubenswrapper[4923]: E0311 13:25:01.768767 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-rdvss" podUID="8bdd5510-3bb1-42f0-8646-c45f11bdeafe" Mar 11 13:25:02 crc kubenswrapper[4923]: I0311 13:25:02.093495 4923 ???:1] "http: TLS handshake error from 192.168.126.11:44202: no serving certificate available for the kubelet" Mar 11 13:25:02 crc kubenswrapper[4923]: I0311 13:25:02.711581 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-57kdl_a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04/kube-multus-additional-cni-plugins/0.log" Mar 11 13:25:02 crc kubenswrapper[4923]: I0311 13:25:02.711923 4923 generic.go:334] "Generic (PLEG): container finished" podID="a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04" containerID="d348f00e8a6e4b07e7a23195b2d70fdf885db23f4ce1b740a3fcb77ccd7756ec" exitCode=137 Mar 11 13:25:02 crc kubenswrapper[4923]: I0311 13:25:02.712036 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" event={"ID":"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04","Type":"ContainerDied","Data":"d348f00e8a6e4b07e7a23195b2d70fdf885db23f4ce1b740a3fcb77ccd7756ec"} Mar 11 13:25:02 crc kubenswrapper[4923]: I0311 13:25:02.753887 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=5.7538642509999995 podStartE2EDuration="5.753864251s" podCreationTimestamp="2026-03-11 13:24:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:25:02.747570518 +0000 UTC m=+150.429472242" watchObservedRunningTime="2026-03-11 13:25:02.753864251 +0000 UTC m=+150.435765965" Mar 11 13:25:03 crc kubenswrapper[4923]: E0311 13:25:03.337587 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-rdvss" podUID="8bdd5510-3bb1-42f0-8646-c45f11bdeafe" Mar 11 13:25:03 crc kubenswrapper[4923]: E0311 13:25:03.337666 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-8f4cb" podUID="52f59385-23ad-4163-a766-b9d21e57ad6c" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.404041 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" Mar 11 13:25:03 crc kubenswrapper[4923]: E0311 13:25:03.406803 4923 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 11 13:25:03 crc kubenswrapper[4923]: E0311 13:25:03.407054 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w48wt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-s9ncx_openshift-marketplace(dbba7a9b-2c9b-4770-ac96-b89390457eab): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 11 13:25:03 crc kubenswrapper[4923]: E0311 13:25:03.408212 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-s9ncx" podUID="dbba7a9b-2c9b-4770-ac96-b89390457eab" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.416760 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87afdf29-c448-483d-9cd5-e6828425c59b-serving-cert\") pod \"87afdf29-c448-483d-9cd5-e6828425c59b\" (UID: \"87afdf29-c448-483d-9cd5-e6828425c59b\") " Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.416810 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/87afdf29-c448-483d-9cd5-e6828425c59b-proxy-ca-bundles\") pod \"87afdf29-c448-483d-9cd5-e6828425c59b\" (UID: \"87afdf29-c448-483d-9cd5-e6828425c59b\") " Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.416844 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94wrw\" (UniqueName: \"kubernetes.io/projected/87afdf29-c448-483d-9cd5-e6828425c59b-kube-api-access-94wrw\") pod \"87afdf29-c448-483d-9cd5-e6828425c59b\" (UID: \"87afdf29-c448-483d-9cd5-e6828425c59b\") " Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.416901 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/87afdf29-c448-483d-9cd5-e6828425c59b-client-ca\") pod \"87afdf29-c448-483d-9cd5-e6828425c59b\" (UID: \"87afdf29-c448-483d-9cd5-e6828425c59b\") " Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.416963 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87afdf29-c448-483d-9cd5-e6828425c59b-config\") pod \"87afdf29-c448-483d-9cd5-e6828425c59b\" (UID: \"87afdf29-c448-483d-9cd5-e6828425c59b\") " Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.418115 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87afdf29-c448-483d-9cd5-e6828425c59b-client-ca" (OuterVolumeSpecName: "client-ca") pod "87afdf29-c448-483d-9cd5-e6828425c59b" (UID: "87afdf29-c448-483d-9cd5-e6828425c59b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.418324 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87afdf29-c448-483d-9cd5-e6828425c59b-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "87afdf29-c448-483d-9cd5-e6828425c59b" (UID: "87afdf29-c448-483d-9cd5-e6828425c59b"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.418392 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87afdf29-c448-483d-9cd5-e6828425c59b-config" (OuterVolumeSpecName: "config") pod "87afdf29-c448-483d-9cd5-e6828425c59b" (UID: "87afdf29-c448-483d-9cd5-e6828425c59b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.425052 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87afdf29-c448-483d-9cd5-e6828425c59b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "87afdf29-c448-483d-9cd5-e6828425c59b" (UID: "87afdf29-c448-483d-9cd5-e6828425c59b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.429095 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87afdf29-c448-483d-9cd5-e6828425c59b-kube-api-access-94wrw" (OuterVolumeSpecName: "kube-api-access-94wrw") pod "87afdf29-c448-483d-9cd5-e6828425c59b" (UID: "87afdf29-c448-483d-9cd5-e6828425c59b"). InnerVolumeSpecName "kube-api-access-94wrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.434095 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-b546dc45b-ps7l8"] Mar 11 13:25:03 crc kubenswrapper[4923]: E0311 13:25:03.434294 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87afdf29-c448-483d-9cd5-e6828425c59b" containerName="controller-manager" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.434306 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="87afdf29-c448-483d-9cd5-e6828425c59b" containerName="controller-manager" Mar 11 13:25:03 crc kubenswrapper[4923]: E0311 13:25:03.434316 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87bd3c68-b0fc-4008-ad02-920658fe0f01" containerName="pruner" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.434321 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="87bd3c68-b0fc-4008-ad02-920658fe0f01" containerName="pruner" Mar 11 13:25:03 crc kubenswrapper[4923]: E0311 13:25:03.434335 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b93878b7-e524-49b0-89af-dfe5de598a5e" containerName="pruner" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.434361 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="b93878b7-e524-49b0-89af-dfe5de598a5e" containerName="pruner" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.434460 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="87bd3c68-b0fc-4008-ad02-920658fe0f01" containerName="pruner" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.434472 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="87afdf29-c448-483d-9cd5-e6828425c59b" containerName="controller-manager" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.434480 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="b93878b7-e524-49b0-89af-dfe5de598a5e" containerName="pruner" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.434905 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.441867 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-b546dc45b-ps7l8"] Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.518774 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5ae7f926-7941-4b70-bfc8-4807cad4411c-proxy-ca-bundles\") pod \"controller-manager-b546dc45b-ps7l8\" (UID: \"5ae7f926-7941-4b70-bfc8-4807cad4411c\") " pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.518823 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ae7f926-7941-4b70-bfc8-4807cad4411c-client-ca\") pod \"controller-manager-b546dc45b-ps7l8\" (UID: \"5ae7f926-7941-4b70-bfc8-4807cad4411c\") " pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.518842 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ae7f926-7941-4b70-bfc8-4807cad4411c-serving-cert\") pod \"controller-manager-b546dc45b-ps7l8\" (UID: \"5ae7f926-7941-4b70-bfc8-4807cad4411c\") " pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.518989 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ae7f926-7941-4b70-bfc8-4807cad4411c-config\") pod \"controller-manager-b546dc45b-ps7l8\" (UID: \"5ae7f926-7941-4b70-bfc8-4807cad4411c\") " pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.519043 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvvbs\" (UniqueName: \"kubernetes.io/projected/5ae7f926-7941-4b70-bfc8-4807cad4411c-kube-api-access-bvvbs\") pod \"controller-manager-b546dc45b-ps7l8\" (UID: \"5ae7f926-7941-4b70-bfc8-4807cad4411c\") " pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.519192 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/87afdf29-c448-483d-9cd5-e6828425c59b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.519203 4923 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/87afdf29-c448-483d-9cd5-e6828425c59b-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.519214 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94wrw\" (UniqueName: \"kubernetes.io/projected/87afdf29-c448-483d-9cd5-e6828425c59b-kube-api-access-94wrw\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.519222 4923 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/87afdf29-c448-483d-9cd5-e6828425c59b-client-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.519231 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/87afdf29-c448-483d-9cd5-e6828425c59b-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.620260 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5ae7f926-7941-4b70-bfc8-4807cad4411c-proxy-ca-bundles\") pod \"controller-manager-b546dc45b-ps7l8\" (UID: \"5ae7f926-7941-4b70-bfc8-4807cad4411c\") " pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.620579 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ae7f926-7941-4b70-bfc8-4807cad4411c-client-ca\") pod \"controller-manager-b546dc45b-ps7l8\" (UID: \"5ae7f926-7941-4b70-bfc8-4807cad4411c\") " pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.620660 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ae7f926-7941-4b70-bfc8-4807cad4411c-serving-cert\") pod \"controller-manager-b546dc45b-ps7l8\" (UID: \"5ae7f926-7941-4b70-bfc8-4807cad4411c\") " pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.620752 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ae7f926-7941-4b70-bfc8-4807cad4411c-config\") pod \"controller-manager-b546dc45b-ps7l8\" (UID: \"5ae7f926-7941-4b70-bfc8-4807cad4411c\") " pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.620832 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvvbs\" (UniqueName: \"kubernetes.io/projected/5ae7f926-7941-4b70-bfc8-4807cad4411c-kube-api-access-bvvbs\") pod \"controller-manager-b546dc45b-ps7l8\" (UID: \"5ae7f926-7941-4b70-bfc8-4807cad4411c\") " pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.621411 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5ae7f926-7941-4b70-bfc8-4807cad4411c-proxy-ca-bundles\") pod \"controller-manager-b546dc45b-ps7l8\" (UID: \"5ae7f926-7941-4b70-bfc8-4807cad4411c\") " pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.621453 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ae7f926-7941-4b70-bfc8-4807cad4411c-client-ca\") pod \"controller-manager-b546dc45b-ps7l8\" (UID: \"5ae7f926-7941-4b70-bfc8-4807cad4411c\") " pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.622256 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ae7f926-7941-4b70-bfc8-4807cad4411c-config\") pod \"controller-manager-b546dc45b-ps7l8\" (UID: \"5ae7f926-7941-4b70-bfc8-4807cad4411c\") " pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.637388 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ae7f926-7941-4b70-bfc8-4807cad4411c-serving-cert\") pod \"controller-manager-b546dc45b-ps7l8\" (UID: \"5ae7f926-7941-4b70-bfc8-4807cad4411c\") " pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.639953 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvvbs\" (UniqueName: \"kubernetes.io/projected/5ae7f926-7941-4b70-bfc8-4807cad4411c-kube-api-access-bvvbs\") pod \"controller-manager-b546dc45b-ps7l8\" (UID: \"5ae7f926-7941-4b70-bfc8-4807cad4411c\") " pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.717883 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" event={"ID":"87afdf29-c448-483d-9cd5-e6828425c59b","Type":"ContainerDied","Data":"4aff2d39fd91159ce2e5e1a1f507339fb01f8928d79a34b62c096b9b3fafcaea"} Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.717937 4923 scope.go:117] "RemoveContainer" containerID="602f5459db5cfde4f8cb7949b672b34d021c30813b2ad553c98879beff34c495" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.718206 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-66d9d6d464-j9rb9" Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.747661 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-66d9d6d464-j9rb9"] Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.750406 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-66d9d6d464-j9rb9"] Mar 11 13:25:03 crc kubenswrapper[4923]: I0311 13:25:03.796360 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" Mar 11 13:25:03 crc kubenswrapper[4923]: E0311 13:25:03.914674 4923 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 11 13:25:03 crc kubenswrapper[4923]: E0311 13:25:03.914814 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8mzqq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-cqdq8_openshift-marketplace(7ce0bb4a-0841-4449-bc02-6c219731272e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 11 13:25:03 crc kubenswrapper[4923]: E0311 13:25:03.916047 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-cqdq8" podUID="7ce0bb4a-0841-4449-bc02-6c219731272e" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.033921 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87afdf29-c448-483d-9cd5-e6828425c59b" path="/var/lib/kubelet/pods/87afdf29-c448-483d-9cd5-e6828425c59b/volumes" Mar 11 13:25:05 crc kubenswrapper[4923]: E0311 13:25:05.331206 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cqdq8" podUID="7ce0bb4a-0841-4449-bc02-6c219731272e" Mar 11 13:25:05 crc kubenswrapper[4923]: E0311 13:25:05.331968 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-s9ncx" podUID="dbba7a9b-2c9b-4770-ac96-b89390457eab" Mar 11 13:25:05 crc kubenswrapper[4923]: E0311 13:25:05.430292 4923 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 11 13:25:05 crc kubenswrapper[4923]: E0311 13:25:05.430750 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hnsfj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-mcgpl_openshift-marketplace(9ab29276-3a67-42c9-8308-2ee71a3dd4dc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 11 13:25:05 crc kubenswrapper[4923]: E0311 13:25:05.431930 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-mcgpl" podUID="9ab29276-3a67-42c9-8308-2ee71a3dd4dc" Mar 11 13:25:05 crc kubenswrapper[4923]: E0311 13:25:05.444550 4923 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 11 13:25:05 crc kubenswrapper[4923]: E0311 13:25:05.444753 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vq82b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-jcncp_openshift-marketplace(ac5c84a5-f129-440d-8423-f8d86b00e3b9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 11 13:25:05 crc kubenswrapper[4923]: E0311 13:25:05.446349 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-jcncp" podUID="ac5c84a5-f129-440d-8423-f8d86b00e3b9" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.472760 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.508203 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-57kdl_a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04/kube-multus-additional-cni-plugins/0.log" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.508279 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.514867 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8"] Mar 11 13:25:05 crc kubenswrapper[4923]: E0311 13:25:05.515377 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04" containerName="kube-multus-additional-cni-plugins" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.515395 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04" containerName="kube-multus-additional-cni-plugins" Mar 11 13:25:05 crc kubenswrapper[4923]: E0311 13:25:05.515408 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8acfb1c-a8e0-4f96-b1ba-7a3210b30560" containerName="route-controller-manager" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.515415 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8acfb1c-a8e0-4f96-b1ba-7a3210b30560" containerName="route-controller-manager" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.515521 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04" containerName="kube-multus-additional-cni-plugins" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.515535 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8acfb1c-a8e0-4f96-b1ba-7a3210b30560" containerName="route-controller-manager" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.515859 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.547774 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8"] Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.550554 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-tuning-conf-dir\") pod \"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04\" (UID: \"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04\") " Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.550627 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-serving-cert\") pod \"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560\" (UID: \"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560\") " Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.550714 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-client-ca\") pod \"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560\" (UID: \"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560\") " Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.550752 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-ready\") pod \"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04\" (UID: \"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04\") " Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.550779 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rj5bs\" (UniqueName: \"kubernetes.io/projected/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-kube-api-access-rj5bs\") pod \"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04\" (UID: \"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04\") " Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.553808 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-config\") pod \"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560\" (UID: \"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560\") " Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.554413 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9kzt\" (UniqueName: \"kubernetes.io/projected/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-kube-api-access-k9kzt\") pod \"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560\" (UID: \"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560\") " Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.554723 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-config" (OuterVolumeSpecName: "config") pod "d8acfb1c-a8e0-4f96-b1ba-7a3210b30560" (UID: "d8acfb1c-a8e0-4f96-b1ba-7a3210b30560"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.554798 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04" (UID: "a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.555160 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-cni-sysctl-allowlist\") pod \"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04\" (UID: \"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04\") " Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.555664 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-client-ca" (OuterVolumeSpecName: "client-ca") pod "d8acfb1c-a8e0-4f96-b1ba-7a3210b30560" (UID: "d8acfb1c-a8e0-4f96-b1ba-7a3210b30560"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.555962 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04" (UID: "a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.556378 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlddn\" (UniqueName: \"kubernetes.io/projected/f5afeb43-6cab-478d-8fdc-fd26e19d6265-kube-api-access-hlddn\") pod \"route-controller-manager-6b978fb8fd-dslf8\" (UID: \"f5afeb43-6cab-478d-8fdc-fd26e19d6265\") " pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.556461 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f5afeb43-6cab-478d-8fdc-fd26e19d6265-client-ca\") pod \"route-controller-manager-6b978fb8fd-dslf8\" (UID: \"f5afeb43-6cab-478d-8fdc-fd26e19d6265\") " pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.556566 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5afeb43-6cab-478d-8fdc-fd26e19d6265-config\") pod \"route-controller-manager-6b978fb8fd-dslf8\" (UID: \"f5afeb43-6cab-478d-8fdc-fd26e19d6265\") " pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.556677 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f5afeb43-6cab-478d-8fdc-fd26e19d6265-serving-cert\") pod \"route-controller-manager-6b978fb8fd-dslf8\" (UID: \"f5afeb43-6cab-478d-8fdc-fd26e19d6265\") " pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.556761 4923 reconciler_common.go:293] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-tuning-conf-dir\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.556773 4923 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-client-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.556788 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.556798 4923 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.557104 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-ready" (OuterVolumeSpecName: "ready") pod "a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04" (UID: "a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.562385 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d8acfb1c-a8e0-4f96-b1ba-7a3210b30560" (UID: "d8acfb1c-a8e0-4f96-b1ba-7a3210b30560"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.562724 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-kube-api-access-rj5bs" (OuterVolumeSpecName: "kube-api-access-rj5bs") pod "a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04" (UID: "a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04"). InnerVolumeSpecName "kube-api-access-rj5bs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.565972 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-kube-api-access-k9kzt" (OuterVolumeSpecName: "kube-api-access-k9kzt") pod "d8acfb1c-a8e0-4f96-b1ba-7a3210b30560" (UID: "d8acfb1c-a8e0-4f96-b1ba-7a3210b30560"). InnerVolumeSpecName "kube-api-access-k9kzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.588436 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-b546dc45b-ps7l8"] Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.599484 4923 patch_prober.go:28] interesting pod/route-controller-manager-56b676fd75-kcjj5 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.48:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.599552 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" podUID="d8acfb1c-a8e0-4f96-b1ba-7a3210b30560" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.48:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.658281 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlddn\" (UniqueName: \"kubernetes.io/projected/f5afeb43-6cab-478d-8fdc-fd26e19d6265-kube-api-access-hlddn\") pod \"route-controller-manager-6b978fb8fd-dslf8\" (UID: \"f5afeb43-6cab-478d-8fdc-fd26e19d6265\") " pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.658377 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f5afeb43-6cab-478d-8fdc-fd26e19d6265-client-ca\") pod \"route-controller-manager-6b978fb8fd-dslf8\" (UID: \"f5afeb43-6cab-478d-8fdc-fd26e19d6265\") " pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.658440 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5afeb43-6cab-478d-8fdc-fd26e19d6265-config\") pod \"route-controller-manager-6b978fb8fd-dslf8\" (UID: \"f5afeb43-6cab-478d-8fdc-fd26e19d6265\") " pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.658495 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f5afeb43-6cab-478d-8fdc-fd26e19d6265-serving-cert\") pod \"route-controller-manager-6b978fb8fd-dslf8\" (UID: \"f5afeb43-6cab-478d-8fdc-fd26e19d6265\") " pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.658549 4923 reconciler_common.go:293] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-ready\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.658564 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rj5bs\" (UniqueName: \"kubernetes.io/projected/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04-kube-api-access-rj5bs\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.658580 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9kzt\" (UniqueName: \"kubernetes.io/projected/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-kube-api-access-k9kzt\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.658693 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.662491 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5afeb43-6cab-478d-8fdc-fd26e19d6265-config\") pod \"route-controller-manager-6b978fb8fd-dslf8\" (UID: \"f5afeb43-6cab-478d-8fdc-fd26e19d6265\") " pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.662519 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f5afeb43-6cab-478d-8fdc-fd26e19d6265-client-ca\") pod \"route-controller-manager-6b978fb8fd-dslf8\" (UID: \"f5afeb43-6cab-478d-8fdc-fd26e19d6265\") " pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.670823 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f5afeb43-6cab-478d-8fdc-fd26e19d6265-serving-cert\") pod \"route-controller-manager-6b978fb8fd-dslf8\" (UID: \"f5afeb43-6cab-478d-8fdc-fd26e19d6265\") " pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.680881 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlddn\" (UniqueName: \"kubernetes.io/projected/f5afeb43-6cab-478d-8fdc-fd26e19d6265-kube-api-access-hlddn\") pod \"route-controller-manager-6b978fb8fd-dslf8\" (UID: \"f5afeb43-6cab-478d-8fdc-fd26e19d6265\") " pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.729655 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnjrh" event={"ID":"094d43b0-8aa7-4acc-9d97-9c00874ee1a8","Type":"ContainerStarted","Data":"786a32861029cbd731746540a5ae3701ac6ca72dc83eff5f0beca46f6037f092"} Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.733243 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-97n98" event={"ID":"08394455-c98e-4e73-b798-3a9c2ff5e708","Type":"ContainerStarted","Data":"f8afffeb3abdc4cefd142f962216f37fbba1758b9046e280c28aef899cb1446f"} Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.734762 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" event={"ID":"5ae7f926-7941-4b70-bfc8-4807cad4411c","Type":"ContainerStarted","Data":"79301436716f4d257926b93895e3de9554c022aa7ab860ba583ce289746d2d32"} Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.734792 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" event={"ID":"5ae7f926-7941-4b70-bfc8-4807cad4411c","Type":"ContainerStarted","Data":"fc77f7619ee8548e38bf963c724b72bcb147311a124fa5b1694c468aa44b9c40"} Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.735425 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.736923 4923 patch_prober.go:28] interesting pod/controller-manager-b546dc45b-ps7l8 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.56:8443/healthz\": dial tcp 10.217.0.56:8443: connect: connection refused" start-of-body= Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.737005 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" podUID="5ae7f926-7941-4b70-bfc8-4807cad4411c" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.56:8443/healthz\": dial tcp 10.217.0.56:8443: connect: connection refused" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.738389 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.739691 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5" event={"ID":"d8acfb1c-a8e0-4f96-b1ba-7a3210b30560","Type":"ContainerDied","Data":"02f3c4af7c39203780fa6b6ae3dda4f4c7aae099077eed500c8de75933f3367f"} Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.739742 4923 scope.go:117] "RemoveContainer" containerID="a22b4179139b9031d2712e5439138ae6419766d8b43c69e76d47e5f30c937b02" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.741940 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-57kdl_a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04/kube-multus-additional-cni-plugins/0.log" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.742065 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" event={"ID":"a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04","Type":"ContainerDied","Data":"1de5ae939254de4f0275f4db1cc697cb33c414e6163743ef45a0461bf150cd47"} Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.742113 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-57kdl" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.763115 4923 scope.go:117] "RemoveContainer" containerID="d348f00e8a6e4b07e7a23195b2d70fdf885db23f4ce1b740a3fcb77ccd7756ec" Mar 11 13:25:05 crc kubenswrapper[4923]: E0311 13:25:05.775080 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-jcncp" podUID="ac5c84a5-f129-440d-8423-f8d86b00e3b9" Mar 11 13:25:05 crc kubenswrapper[4923]: E0311 13:25:05.775087 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-mcgpl" podUID="9ab29276-3a67-42c9-8308-2ee71a3dd4dc" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.824198 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" podStartSLOduration=13.824182257 podStartE2EDuration="13.824182257s" podCreationTimestamp="2026-03-11 13:24:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:25:05.811588351 +0000 UTC m=+153.493490075" watchObservedRunningTime="2026-03-11 13:25:05.824182257 +0000 UTC m=+153.506083971" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.827110 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5"] Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.834273 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56b676fd75-kcjj5"] Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.842496 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-57kdl"] Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.847621 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-57kdl"] Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.847683 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" Mar 11 13:25:05 crc kubenswrapper[4923]: I0311 13:25:05.862940 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 11 13:25:06 crc kubenswrapper[4923]: I0311 13:25:06.114743 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8"] Mar 11 13:25:06 crc kubenswrapper[4923]: W0311 13:25:06.134114 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5afeb43_6cab_478d_8fdc_fd26e19d6265.slice/crio-f0a851cd894507af119ab7237ecfd450038d34f913808da237cb6dfdd77f316e WatchSource:0}: Error finding container f0a851cd894507af119ab7237ecfd450038d34f913808da237cb6dfdd77f316e: Status 404 returned error can't find the container with id f0a851cd894507af119ab7237ecfd450038d34f913808da237cb6dfdd77f316e Mar 11 13:25:06 crc kubenswrapper[4923]: I0311 13:25:06.753012 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-h59nb" Mar 11 13:25:06 crc kubenswrapper[4923]: I0311 13:25:06.783163 4923 generic.go:334] "Generic (PLEG): container finished" podID="094d43b0-8aa7-4acc-9d97-9c00874ee1a8" containerID="786a32861029cbd731746540a5ae3701ac6ca72dc83eff5f0beca46f6037f092" exitCode=0 Mar 11 13:25:06 crc kubenswrapper[4923]: I0311 13:25:06.783243 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnjrh" event={"ID":"094d43b0-8aa7-4acc-9d97-9c00874ee1a8","Type":"ContainerDied","Data":"786a32861029cbd731746540a5ae3701ac6ca72dc83eff5f0beca46f6037f092"} Mar 11 13:25:06 crc kubenswrapper[4923]: I0311 13:25:06.788611 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" event={"ID":"f5afeb43-6cab-478d-8fdc-fd26e19d6265","Type":"ContainerStarted","Data":"6c59068ddb1058e51f5086173a2d36edeae402f4d426bcb3cbe3edd739137d90"} Mar 11 13:25:06 crc kubenswrapper[4923]: I0311 13:25:06.788666 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" event={"ID":"f5afeb43-6cab-478d-8fdc-fd26e19d6265","Type":"ContainerStarted","Data":"f0a851cd894507af119ab7237ecfd450038d34f913808da237cb6dfdd77f316e"} Mar 11 13:25:06 crc kubenswrapper[4923]: I0311 13:25:06.789393 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" Mar 11 13:25:06 crc kubenswrapper[4923]: I0311 13:25:06.798033 4923 generic.go:334] "Generic (PLEG): container finished" podID="08394455-c98e-4e73-b798-3a9c2ff5e708" containerID="f8afffeb3abdc4cefd142f962216f37fbba1758b9046e280c28aef899cb1446f" exitCode=0 Mar 11 13:25:06 crc kubenswrapper[4923]: I0311 13:25:06.798152 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-97n98" event={"ID":"08394455-c98e-4e73-b798-3a9c2ff5e708","Type":"ContainerDied","Data":"f8afffeb3abdc4cefd142f962216f37fbba1758b9046e280c28aef899cb1446f"} Mar 11 13:25:06 crc kubenswrapper[4923]: I0311 13:25:06.809063 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" Mar 11 13:25:06 crc kubenswrapper[4923]: I0311 13:25:06.824595 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" Mar 11 13:25:06 crc kubenswrapper[4923]: I0311 13:25:06.840021 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" podStartSLOduration=14.839999773 podStartE2EDuration="14.839999773s" podCreationTimestamp="2026-03-11 13:24:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:25:06.837564556 +0000 UTC m=+154.519466270" watchObservedRunningTime="2026-03-11 13:25:06.839999773 +0000 UTC m=+154.521901497" Mar 11 13:25:07 crc kubenswrapper[4923]: I0311 13:25:07.029123 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04" path="/var/lib/kubelet/pods/a7cbdd24-fc6a-4194-9d8a-eb87e5b4aa04/volumes" Mar 11 13:25:07 crc kubenswrapper[4923]: I0311 13:25:07.030180 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8acfb1c-a8e0-4f96-b1ba-7a3210b30560" path="/var/lib/kubelet/pods/d8acfb1c-a8e0-4f96-b1ba-7a3210b30560/volumes" Mar 11 13:25:08 crc kubenswrapper[4923]: I0311 13:25:08.752826 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 11 13:25:08 crc kubenswrapper[4923]: I0311 13:25:08.754213 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 11 13:25:08 crc kubenswrapper[4923]: I0311 13:25:08.757996 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 11 13:25:08 crc kubenswrapper[4923]: I0311 13:25:08.761411 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 11 13:25:08 crc kubenswrapper[4923]: I0311 13:25:08.762068 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 11 13:25:08 crc kubenswrapper[4923]: I0311 13:25:08.803590 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66f3599b-8971-44c8-8fcd-63f92c7cbaea-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"66f3599b-8971-44c8-8fcd-63f92c7cbaea\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 11 13:25:08 crc kubenswrapper[4923]: I0311 13:25:08.803656 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/66f3599b-8971-44c8-8fcd-63f92c7cbaea-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"66f3599b-8971-44c8-8fcd-63f92c7cbaea\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 11 13:25:08 crc kubenswrapper[4923]: I0311 13:25:08.818058 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnjrh" event={"ID":"094d43b0-8aa7-4acc-9d97-9c00874ee1a8","Type":"ContainerStarted","Data":"fe4253ba3094d50d1f38cf57ba1844808af2c64277e0140ca9e55de3afdd26a8"} Mar 11 13:25:08 crc kubenswrapper[4923]: I0311 13:25:08.823059 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-97n98" event={"ID":"08394455-c98e-4e73-b798-3a9c2ff5e708","Type":"ContainerStarted","Data":"6e5fc33d552910006afd99bf5fbbf246fc57409a0d4cd61463d2d4417b5887db"} Mar 11 13:25:08 crc kubenswrapper[4923]: I0311 13:25:08.844434 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rnjrh" podStartSLOduration=3.1194935 podStartE2EDuration="33.844421089s" podCreationTimestamp="2026-03-11 13:24:35 +0000 UTC" firstStartedPulling="2026-03-11 13:24:37.523075857 +0000 UTC m=+125.204977571" lastFinishedPulling="2026-03-11 13:25:08.248003446 +0000 UTC m=+155.929905160" observedRunningTime="2026-03-11 13:25:08.841637662 +0000 UTC m=+156.523539376" watchObservedRunningTime="2026-03-11 13:25:08.844421089 +0000 UTC m=+156.526322803" Mar 11 13:25:08 crc kubenswrapper[4923]: I0311 13:25:08.863348 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-97n98" podStartSLOduration=4.368205837 podStartE2EDuration="33.863320928s" podCreationTimestamp="2026-03-11 13:24:35 +0000 UTC" firstStartedPulling="2026-03-11 13:24:38.547532138 +0000 UTC m=+126.229433852" lastFinishedPulling="2026-03-11 13:25:08.042647229 +0000 UTC m=+155.724548943" observedRunningTime="2026-03-11 13:25:08.860333625 +0000 UTC m=+156.542235329" watchObservedRunningTime="2026-03-11 13:25:08.863320928 +0000 UTC m=+156.545222642" Mar 11 13:25:08 crc kubenswrapper[4923]: I0311 13:25:08.905011 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66f3599b-8971-44c8-8fcd-63f92c7cbaea-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"66f3599b-8971-44c8-8fcd-63f92c7cbaea\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 11 13:25:08 crc kubenswrapper[4923]: I0311 13:25:08.905069 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/66f3599b-8971-44c8-8fcd-63f92c7cbaea-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"66f3599b-8971-44c8-8fcd-63f92c7cbaea\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 11 13:25:08 crc kubenswrapper[4923]: I0311 13:25:08.905202 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/66f3599b-8971-44c8-8fcd-63f92c7cbaea-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"66f3599b-8971-44c8-8fcd-63f92c7cbaea\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 11 13:25:08 crc kubenswrapper[4923]: I0311 13:25:08.923774 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66f3599b-8971-44c8-8fcd-63f92c7cbaea-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"66f3599b-8971-44c8-8fcd-63f92c7cbaea\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 11 13:25:09 crc kubenswrapper[4923]: I0311 13:25:09.069198 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 11 13:25:09 crc kubenswrapper[4923]: I0311 13:25:09.475582 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 11 13:25:09 crc kubenswrapper[4923]: I0311 13:25:09.853822 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"66f3599b-8971-44c8-8fcd-63f92c7cbaea","Type":"ContainerStarted","Data":"5b16246d907c22ce841d5e551c663a3bb42ea97f35322019e5d21ce80ef4a38f"} Mar 11 13:25:10 crc kubenswrapper[4923]: I0311 13:25:10.860997 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"66f3599b-8971-44c8-8fcd-63f92c7cbaea","Type":"ContainerStarted","Data":"d801fc45325e77f8cbb8e845a6d58e70c1c94093c95644e0068a5f6628cb5fae"} Mar 11 13:25:11 crc kubenswrapper[4923]: I0311 13:25:11.866858 4923 generic.go:334] "Generic (PLEG): container finished" podID="66f3599b-8971-44c8-8fcd-63f92c7cbaea" containerID="d801fc45325e77f8cbb8e845a6d58e70c1c94093c95644e0068a5f6628cb5fae" exitCode=0 Mar 11 13:25:11 crc kubenswrapper[4923]: I0311 13:25:11.867154 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"66f3599b-8971-44c8-8fcd-63f92c7cbaea","Type":"ContainerDied","Data":"d801fc45325e77f8cbb8e845a6d58e70c1c94093c95644e0068a5f6628cb5fae"} Mar 11 13:25:12 crc kubenswrapper[4923]: I0311 13:25:12.085836 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-b546dc45b-ps7l8"] Mar 11 13:25:12 crc kubenswrapper[4923]: I0311 13:25:12.086039 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" podUID="5ae7f926-7941-4b70-bfc8-4807cad4411c" containerName="controller-manager" containerID="cri-o://79301436716f4d257926b93895e3de9554c022aa7ab860ba583ce289746d2d32" gracePeriod=30 Mar 11 13:25:12 crc kubenswrapper[4923]: I0311 13:25:12.177638 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8"] Mar 11 13:25:12 crc kubenswrapper[4923]: I0311 13:25:12.177830 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" podUID="f5afeb43-6cab-478d-8fdc-fd26e19d6265" containerName="route-controller-manager" containerID="cri-o://6c59068ddb1058e51f5086173a2d36edeae402f4d426bcb3cbe3edd739137d90" gracePeriod=30 Mar 11 13:25:12 crc kubenswrapper[4923]: I0311 13:25:12.874709 4923 generic.go:334] "Generic (PLEG): container finished" podID="f5afeb43-6cab-478d-8fdc-fd26e19d6265" containerID="6c59068ddb1058e51f5086173a2d36edeae402f4d426bcb3cbe3edd739137d90" exitCode=0 Mar 11 13:25:12 crc kubenswrapper[4923]: I0311 13:25:12.874791 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" event={"ID":"f5afeb43-6cab-478d-8fdc-fd26e19d6265","Type":"ContainerDied","Data":"6c59068ddb1058e51f5086173a2d36edeae402f4d426bcb3cbe3edd739137d90"} Mar 11 13:25:12 crc kubenswrapper[4923]: I0311 13:25:12.877628 4923 generic.go:334] "Generic (PLEG): container finished" podID="5ae7f926-7941-4b70-bfc8-4807cad4411c" containerID="79301436716f4d257926b93895e3de9554c022aa7ab860ba583ce289746d2d32" exitCode=0 Mar 11 13:25:12 crc kubenswrapper[4923]: I0311 13:25:12.877727 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" event={"ID":"5ae7f926-7941-4b70-bfc8-4807cad4411c","Type":"ContainerDied","Data":"79301436716f4d257926b93895e3de9554c022aa7ab860ba583ce289746d2d32"} Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.189711 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.206878 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.264667 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ae7f926-7941-4b70-bfc8-4807cad4411c-config\") pod \"5ae7f926-7941-4b70-bfc8-4807cad4411c\" (UID: \"5ae7f926-7941-4b70-bfc8-4807cad4411c\") " Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.264703 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ae7f926-7941-4b70-bfc8-4807cad4411c-serving-cert\") pod \"5ae7f926-7941-4b70-bfc8-4807cad4411c\" (UID: \"5ae7f926-7941-4b70-bfc8-4807cad4411c\") " Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.264765 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/66f3599b-8971-44c8-8fcd-63f92c7cbaea-kubelet-dir\") pod \"66f3599b-8971-44c8-8fcd-63f92c7cbaea\" (UID: \"66f3599b-8971-44c8-8fcd-63f92c7cbaea\") " Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.264793 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66f3599b-8971-44c8-8fcd-63f92c7cbaea-kube-api-access\") pod \"66f3599b-8971-44c8-8fcd-63f92c7cbaea\" (UID: \"66f3599b-8971-44c8-8fcd-63f92c7cbaea\") " Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.264841 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvvbs\" (UniqueName: \"kubernetes.io/projected/5ae7f926-7941-4b70-bfc8-4807cad4411c-kube-api-access-bvvbs\") pod \"5ae7f926-7941-4b70-bfc8-4807cad4411c\" (UID: \"5ae7f926-7941-4b70-bfc8-4807cad4411c\") " Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.264902 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5ae7f926-7941-4b70-bfc8-4807cad4411c-proxy-ca-bundles\") pod \"5ae7f926-7941-4b70-bfc8-4807cad4411c\" (UID: \"5ae7f926-7941-4b70-bfc8-4807cad4411c\") " Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.264925 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ae7f926-7941-4b70-bfc8-4807cad4411c-client-ca\") pod \"5ae7f926-7941-4b70-bfc8-4807cad4411c\" (UID: \"5ae7f926-7941-4b70-bfc8-4807cad4411c\") " Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.265477 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/66f3599b-8971-44c8-8fcd-63f92c7cbaea-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "66f3599b-8971-44c8-8fcd-63f92c7cbaea" (UID: "66f3599b-8971-44c8-8fcd-63f92c7cbaea"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.265677 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.266130 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ae7f926-7941-4b70-bfc8-4807cad4411c-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "5ae7f926-7941-4b70-bfc8-4807cad4411c" (UID: "5ae7f926-7941-4b70-bfc8-4807cad4411c"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.266402 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ae7f926-7941-4b70-bfc8-4807cad4411c-config" (OuterVolumeSpecName: "config") pod "5ae7f926-7941-4b70-bfc8-4807cad4411c" (UID: "5ae7f926-7941-4b70-bfc8-4807cad4411c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.266699 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ae7f926-7941-4b70-bfc8-4807cad4411c-client-ca" (OuterVolumeSpecName: "client-ca") pod "5ae7f926-7941-4b70-bfc8-4807cad4411c" (UID: "5ae7f926-7941-4b70-bfc8-4807cad4411c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.272263 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ae7f926-7941-4b70-bfc8-4807cad4411c-kube-api-access-bvvbs" (OuterVolumeSpecName: "kube-api-access-bvvbs") pod "5ae7f926-7941-4b70-bfc8-4807cad4411c" (UID: "5ae7f926-7941-4b70-bfc8-4807cad4411c"). InnerVolumeSpecName "kube-api-access-bvvbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.272326 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66f3599b-8971-44c8-8fcd-63f92c7cbaea-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "66f3599b-8971-44c8-8fcd-63f92c7cbaea" (UID: "66f3599b-8971-44c8-8fcd-63f92c7cbaea"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.272611 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ae7f926-7941-4b70-bfc8-4807cad4411c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5ae7f926-7941-4b70-bfc8-4807cad4411c" (UID: "5ae7f926-7941-4b70-bfc8-4807cad4411c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.366173 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f5afeb43-6cab-478d-8fdc-fd26e19d6265-serving-cert\") pod \"f5afeb43-6cab-478d-8fdc-fd26e19d6265\" (UID: \"f5afeb43-6cab-478d-8fdc-fd26e19d6265\") " Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.366310 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f5afeb43-6cab-478d-8fdc-fd26e19d6265-client-ca\") pod \"f5afeb43-6cab-478d-8fdc-fd26e19d6265\" (UID: \"f5afeb43-6cab-478d-8fdc-fd26e19d6265\") " Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.366382 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlddn\" (UniqueName: \"kubernetes.io/projected/f5afeb43-6cab-478d-8fdc-fd26e19d6265-kube-api-access-hlddn\") pod \"f5afeb43-6cab-478d-8fdc-fd26e19d6265\" (UID: \"f5afeb43-6cab-478d-8fdc-fd26e19d6265\") " Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.366423 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5afeb43-6cab-478d-8fdc-fd26e19d6265-config\") pod \"f5afeb43-6cab-478d-8fdc-fd26e19d6265\" (UID: \"f5afeb43-6cab-478d-8fdc-fd26e19d6265\") " Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.366650 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/66f3599b-8971-44c8-8fcd-63f92c7cbaea-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.366661 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvvbs\" (UniqueName: \"kubernetes.io/projected/5ae7f926-7941-4b70-bfc8-4807cad4411c-kube-api-access-bvvbs\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.366673 4923 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/5ae7f926-7941-4b70-bfc8-4807cad4411c-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.366681 4923 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5ae7f926-7941-4b70-bfc8-4807cad4411c-client-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.366690 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5ae7f926-7941-4b70-bfc8-4807cad4411c-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.366698 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5ae7f926-7941-4b70-bfc8-4807cad4411c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.366708 4923 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/66f3599b-8971-44c8-8fcd-63f92c7cbaea-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.367283 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5afeb43-6cab-478d-8fdc-fd26e19d6265-config" (OuterVolumeSpecName: "config") pod "f5afeb43-6cab-478d-8fdc-fd26e19d6265" (UID: "f5afeb43-6cab-478d-8fdc-fd26e19d6265"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.367521 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5afeb43-6cab-478d-8fdc-fd26e19d6265-client-ca" (OuterVolumeSpecName: "client-ca") pod "f5afeb43-6cab-478d-8fdc-fd26e19d6265" (UID: "f5afeb43-6cab-478d-8fdc-fd26e19d6265"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.370136 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5afeb43-6cab-478d-8fdc-fd26e19d6265-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f5afeb43-6cab-478d-8fdc-fd26e19d6265" (UID: "f5afeb43-6cab-478d-8fdc-fd26e19d6265"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.373534 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5afeb43-6cab-478d-8fdc-fd26e19d6265-kube-api-access-hlddn" (OuterVolumeSpecName: "kube-api-access-hlddn") pod "f5afeb43-6cab-478d-8fdc-fd26e19d6265" (UID: "f5afeb43-6cab-478d-8fdc-fd26e19d6265"). InnerVolumeSpecName "kube-api-access-hlddn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.467767 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlddn\" (UniqueName: \"kubernetes.io/projected/f5afeb43-6cab-478d-8fdc-fd26e19d6265-kube-api-access-hlddn\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.467806 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f5afeb43-6cab-478d-8fdc-fd26e19d6265-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.467816 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f5afeb43-6cab-478d-8fdc-fd26e19d6265-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.467830 4923 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f5afeb43-6cab-478d-8fdc-fd26e19d6265-client-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.505270 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk"] Mar 11 13:25:13 crc kubenswrapper[4923]: E0311 13:25:13.505546 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66f3599b-8971-44c8-8fcd-63f92c7cbaea" containerName="pruner" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.505568 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="66f3599b-8971-44c8-8fcd-63f92c7cbaea" containerName="pruner" Mar 11 13:25:13 crc kubenswrapper[4923]: E0311 13:25:13.505579 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ae7f926-7941-4b70-bfc8-4807cad4411c" containerName="controller-manager" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.505587 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ae7f926-7941-4b70-bfc8-4807cad4411c" containerName="controller-manager" Mar 11 13:25:13 crc kubenswrapper[4923]: E0311 13:25:13.505596 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5afeb43-6cab-478d-8fdc-fd26e19d6265" containerName="route-controller-manager" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.505603 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5afeb43-6cab-478d-8fdc-fd26e19d6265" containerName="route-controller-manager" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.505704 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5afeb43-6cab-478d-8fdc-fd26e19d6265" containerName="route-controller-manager" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.505719 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ae7f926-7941-4b70-bfc8-4807cad4411c" containerName="controller-manager" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.505727 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="66f3599b-8971-44c8-8fcd-63f92c7cbaea" containerName="pruner" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.506094 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.523337 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk"] Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.568870 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13ae0f7e-b016-4044-b6db-4210a79854ab-config\") pod \"route-controller-manager-5b758c8ff-btrzk\" (UID: \"13ae0f7e-b016-4044-b6db-4210a79854ab\") " pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.568933 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13ae0f7e-b016-4044-b6db-4210a79854ab-client-ca\") pod \"route-controller-manager-5b758c8ff-btrzk\" (UID: \"13ae0f7e-b016-4044-b6db-4210a79854ab\") " pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.568971 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2zwt\" (UniqueName: \"kubernetes.io/projected/13ae0f7e-b016-4044-b6db-4210a79854ab-kube-api-access-b2zwt\") pod \"route-controller-manager-5b758c8ff-btrzk\" (UID: \"13ae0f7e-b016-4044-b6db-4210a79854ab\") " pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.569002 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13ae0f7e-b016-4044-b6db-4210a79854ab-serving-cert\") pod \"route-controller-manager-5b758c8ff-btrzk\" (UID: \"13ae0f7e-b016-4044-b6db-4210a79854ab\") " pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.670473 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13ae0f7e-b016-4044-b6db-4210a79854ab-serving-cert\") pod \"route-controller-manager-5b758c8ff-btrzk\" (UID: \"13ae0f7e-b016-4044-b6db-4210a79854ab\") " pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.670534 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13ae0f7e-b016-4044-b6db-4210a79854ab-config\") pod \"route-controller-manager-5b758c8ff-btrzk\" (UID: \"13ae0f7e-b016-4044-b6db-4210a79854ab\") " pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.670597 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13ae0f7e-b016-4044-b6db-4210a79854ab-client-ca\") pod \"route-controller-manager-5b758c8ff-btrzk\" (UID: \"13ae0f7e-b016-4044-b6db-4210a79854ab\") " pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.670646 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2zwt\" (UniqueName: \"kubernetes.io/projected/13ae0f7e-b016-4044-b6db-4210a79854ab-kube-api-access-b2zwt\") pod \"route-controller-manager-5b758c8ff-btrzk\" (UID: \"13ae0f7e-b016-4044-b6db-4210a79854ab\") " pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.673209 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13ae0f7e-b016-4044-b6db-4210a79854ab-client-ca\") pod \"route-controller-manager-5b758c8ff-btrzk\" (UID: \"13ae0f7e-b016-4044-b6db-4210a79854ab\") " pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.675564 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13ae0f7e-b016-4044-b6db-4210a79854ab-config\") pod \"route-controller-manager-5b758c8ff-btrzk\" (UID: \"13ae0f7e-b016-4044-b6db-4210a79854ab\") " pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.676884 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13ae0f7e-b016-4044-b6db-4210a79854ab-serving-cert\") pod \"route-controller-manager-5b758c8ff-btrzk\" (UID: \"13ae0f7e-b016-4044-b6db-4210a79854ab\") " pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.694664 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2zwt\" (UniqueName: \"kubernetes.io/projected/13ae0f7e-b016-4044-b6db-4210a79854ab-kube-api-access-b2zwt\") pod \"route-controller-manager-5b758c8ff-btrzk\" (UID: \"13ae0f7e-b016-4044-b6db-4210a79854ab\") " pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.819600 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.885460 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"66f3599b-8971-44c8-8fcd-63f92c7cbaea","Type":"ContainerDied","Data":"5b16246d907c22ce841d5e551c663a3bb42ea97f35322019e5d21ce80ef4a38f"} Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.885508 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b16246d907c22ce841d5e551c663a3bb42ea97f35322019e5d21ce80ef4a38f" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.885580 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.897420 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" event={"ID":"f5afeb43-6cab-478d-8fdc-fd26e19d6265","Type":"ContainerDied","Data":"f0a851cd894507af119ab7237ecfd450038d34f913808da237cb6dfdd77f316e"} Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.897460 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.897470 4923 scope.go:117] "RemoveContainer" containerID="6c59068ddb1058e51f5086173a2d36edeae402f4d426bcb3cbe3edd739137d90" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.899718 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" event={"ID":"5ae7f926-7941-4b70-bfc8-4807cad4411c","Type":"ContainerDied","Data":"fc77f7619ee8548e38bf963c724b72bcb147311a124fa5b1694c468aa44b9c40"} Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.899766 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b546dc45b-ps7l8" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.918443 4923 scope.go:117] "RemoveContainer" containerID="79301436716f4d257926b93895e3de9554c022aa7ab860ba583ce289746d2d32" Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.972906 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8"] Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.977627 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6b978fb8fd-dslf8"] Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.985481 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-b546dc45b-ps7l8"] Mar 11 13:25:13 crc kubenswrapper[4923]: I0311 13:25:13.988544 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-b546dc45b-ps7l8"] Mar 11 13:25:14 crc kubenswrapper[4923]: E0311 13:25:14.005528 4923 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5afeb43_6cab_478d_8fdc_fd26e19d6265.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ae7f926_7941_4b70_bfc8_4807cad4411c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ae7f926_7941_4b70_bfc8_4807cad4411c.slice/crio-fc77f7619ee8548e38bf963c724b72bcb147311a124fa5b1694c468aa44b9c40\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-pod66f3599b_8971_44c8_8fcd_63f92c7cbaea.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-pod66f3599b_8971_44c8_8fcd_63f92c7cbaea.slice/crio-5b16246d907c22ce841d5e551c663a3bb42ea97f35322019e5d21ce80ef4a38f\": RecentStats: unable to find data in memory cache]" Mar 11 13:25:14 crc kubenswrapper[4923]: I0311 13:25:14.281936 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk"] Mar 11 13:25:14 crc kubenswrapper[4923]: W0311 13:25:14.291717 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13ae0f7e_b016_4044_b6db_4210a79854ab.slice/crio-57d2ae1d56b993a2af0f1e6a19ae7758401ceef3b2150b6a2f6b57c1c5a8e349 WatchSource:0}: Error finding container 57d2ae1d56b993a2af0f1e6a19ae7758401ceef3b2150b6a2f6b57c1c5a8e349: Status 404 returned error can't find the container with id 57d2ae1d56b993a2af0f1e6a19ae7758401ceef3b2150b6a2f6b57c1c5a8e349 Mar 11 13:25:14 crc kubenswrapper[4923]: I0311 13:25:14.909443 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" event={"ID":"13ae0f7e-b016-4044-b6db-4210a79854ab","Type":"ContainerStarted","Data":"b7ad9e1425218d639b26e21c5bcb52eab894a864e3e4781633b36fb82ba8826f"} Mar 11 13:25:14 crc kubenswrapper[4923]: I0311 13:25:14.909727 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" event={"ID":"13ae0f7e-b016-4044-b6db-4210a79854ab","Type":"ContainerStarted","Data":"57d2ae1d56b993a2af0f1e6a19ae7758401ceef3b2150b6a2f6b57c1c5a8e349"} Mar 11 13:25:14 crc kubenswrapper[4923]: I0311 13:25:14.909882 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" Mar 11 13:25:14 crc kubenswrapper[4923]: I0311 13:25:14.942947 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" Mar 11 13:25:14 crc kubenswrapper[4923]: I0311 13:25:14.961333 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" podStartSLOduration=2.9613183789999997 podStartE2EDuration="2.961318379s" podCreationTimestamp="2026-03-11 13:25:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:25:14.932742275 +0000 UTC m=+162.614643989" watchObservedRunningTime="2026-03-11 13:25:14.961318379 +0000 UTC m=+162.643220093" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.039928 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ae7f926-7941-4b70-bfc8-4807cad4411c" path="/var/lib/kubelet/pods/5ae7f926-7941-4b70-bfc8-4807cad4411c/volumes" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.040569 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5afeb43-6cab-478d-8fdc-fd26e19d6265" path="/var/lib/kubelet/pods/f5afeb43-6cab-478d-8fdc-fd26e19d6265/volumes" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.234304 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hgkc2"] Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.502732 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7f5b598bff-5r5gv"] Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.503360 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.507067 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.507959 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.508712 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.508732 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.508765 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.514945 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7f5b598bff-5r5gv"] Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.516649 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.518786 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.592861 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1100409d-07f5-424b-8b9d-f8e8bd066a5c-client-ca\") pod \"controller-manager-7f5b598bff-5r5gv\" (UID: \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\") " pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.592911 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1100409d-07f5-424b-8b9d-f8e8bd066a5c-config\") pod \"controller-manager-7f5b598bff-5r5gv\" (UID: \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\") " pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.592966 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpw2r\" (UniqueName: \"kubernetes.io/projected/1100409d-07f5-424b-8b9d-f8e8bd066a5c-kube-api-access-zpw2r\") pod \"controller-manager-7f5b598bff-5r5gv\" (UID: \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\") " pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.593051 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1100409d-07f5-424b-8b9d-f8e8bd066a5c-serving-cert\") pod \"controller-manager-7f5b598bff-5r5gv\" (UID: \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\") " pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.593074 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1100409d-07f5-424b-8b9d-f8e8bd066a5c-proxy-ca-bundles\") pod \"controller-manager-7f5b598bff-5r5gv\" (UID: \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\") " pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.693936 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1100409d-07f5-424b-8b9d-f8e8bd066a5c-serving-cert\") pod \"controller-manager-7f5b598bff-5r5gv\" (UID: \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\") " pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.693985 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1100409d-07f5-424b-8b9d-f8e8bd066a5c-proxy-ca-bundles\") pod \"controller-manager-7f5b598bff-5r5gv\" (UID: \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\") " pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.694038 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1100409d-07f5-424b-8b9d-f8e8bd066a5c-client-ca\") pod \"controller-manager-7f5b598bff-5r5gv\" (UID: \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\") " pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.694060 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1100409d-07f5-424b-8b9d-f8e8bd066a5c-config\") pod \"controller-manager-7f5b598bff-5r5gv\" (UID: \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\") " pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.694088 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpw2r\" (UniqueName: \"kubernetes.io/projected/1100409d-07f5-424b-8b9d-f8e8bd066a5c-kube-api-access-zpw2r\") pod \"controller-manager-7f5b598bff-5r5gv\" (UID: \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\") " pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.695303 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1100409d-07f5-424b-8b9d-f8e8bd066a5c-client-ca\") pod \"controller-manager-7f5b598bff-5r5gv\" (UID: \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\") " pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.695478 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1100409d-07f5-424b-8b9d-f8e8bd066a5c-proxy-ca-bundles\") pod \"controller-manager-7f5b598bff-5r5gv\" (UID: \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\") " pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.695916 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1100409d-07f5-424b-8b9d-f8e8bd066a5c-config\") pod \"controller-manager-7f5b598bff-5r5gv\" (UID: \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\") " pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.701434 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1100409d-07f5-424b-8b9d-f8e8bd066a5c-serving-cert\") pod \"controller-manager-7f5b598bff-5r5gv\" (UID: \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\") " pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.712615 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpw2r\" (UniqueName: \"kubernetes.io/projected/1100409d-07f5-424b-8b9d-f8e8bd066a5c-kube-api-access-zpw2r\") pod \"controller-manager-7f5b598bff-5r5gv\" (UID: \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\") " pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.818627 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.870557 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rnjrh" Mar 11 13:25:15 crc kubenswrapper[4923]: I0311 13:25:15.870599 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rnjrh" Mar 11 13:25:16 crc kubenswrapper[4923]: I0311 13:25:16.020094 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rnjrh" Mar 11 13:25:16 crc kubenswrapper[4923]: I0311 13:25:16.071367 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rnjrh" Mar 11 13:25:16 crc kubenswrapper[4923]: I0311 13:25:16.286768 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7f5b598bff-5r5gv"] Mar 11 13:25:16 crc kubenswrapper[4923]: W0311 13:25:16.287743 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1100409d_07f5_424b_8b9d_f8e8bd066a5c.slice/crio-d2985020594cc12b88451cd0b8b01be331706c399d0533c5f0ee1b0bb6bb4ba2 WatchSource:0}: Error finding container d2985020594cc12b88451cd0b8b01be331706c399d0533c5f0ee1b0bb6bb4ba2: Status 404 returned error can't find the container with id d2985020594cc12b88451cd0b8b01be331706c399d0533c5f0ee1b0bb6bb4ba2 Mar 11 13:25:16 crc kubenswrapper[4923]: I0311 13:25:16.568834 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-97n98" Mar 11 13:25:16 crc kubenswrapper[4923]: I0311 13:25:16.569175 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-97n98" Mar 11 13:25:16 crc kubenswrapper[4923]: I0311 13:25:16.630003 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-97n98" Mar 11 13:25:16 crc kubenswrapper[4923]: I0311 13:25:16.920010 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" event={"ID":"1100409d-07f5-424b-8b9d-f8e8bd066a5c","Type":"ContainerStarted","Data":"04ee7610f6b150a515d1dcae9289d9fcb3dcc44708dffead179693c530f02725"} Mar 11 13:25:16 crc kubenswrapper[4923]: I0311 13:25:16.920063 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" event={"ID":"1100409d-07f5-424b-8b9d-f8e8bd066a5c","Type":"ContainerStarted","Data":"d2985020594cc12b88451cd0b8b01be331706c399d0533c5f0ee1b0bb6bb4ba2"} Mar 11 13:25:16 crc kubenswrapper[4923]: I0311 13:25:16.920170 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" Mar 11 13:25:16 crc kubenswrapper[4923]: I0311 13:25:16.922132 4923 generic.go:334] "Generic (PLEG): container finished" podID="8bdd5510-3bb1-42f0-8646-c45f11bdeafe" containerID="436dc8cf923342891f9bf2a15c47e721e0f5b23f0b3e92e7bea7239d8a46ff47" exitCode=0 Mar 11 13:25:16 crc kubenswrapper[4923]: I0311 13:25:16.922222 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdvss" event={"ID":"8bdd5510-3bb1-42f0-8646-c45f11bdeafe","Type":"ContainerDied","Data":"436dc8cf923342891f9bf2a15c47e721e0f5b23f0b3e92e7bea7239d8a46ff47"} Mar 11 13:25:16 crc kubenswrapper[4923]: I0311 13:25:16.926360 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" Mar 11 13:25:16 crc kubenswrapper[4923]: I0311 13:25:16.938712 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" podStartSLOduration=4.938693922 podStartE2EDuration="4.938693922s" podCreationTimestamp="2026-03-11 13:25:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:25:16.934887258 +0000 UTC m=+164.616788972" watchObservedRunningTime="2026-03-11 13:25:16.938693922 +0000 UTC m=+164.620595646" Mar 11 13:25:16 crc kubenswrapper[4923]: I0311 13:25:16.975021 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-97n98" Mar 11 13:25:17 crc kubenswrapper[4923]: I0311 13:25:17.466770 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-97n98"] Mar 11 13:25:17 crc kubenswrapper[4923]: I0311 13:25:17.930698 4923 generic.go:334] "Generic (PLEG): container finished" podID="dbba7a9b-2c9b-4770-ac96-b89390457eab" containerID="17667fa4b595d4d5f150d54c90b0674248035cd9f0def0674ea10cfc58e20fa2" exitCode=0 Mar 11 13:25:17 crc kubenswrapper[4923]: I0311 13:25:17.930745 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s9ncx" event={"ID":"dbba7a9b-2c9b-4770-ac96-b89390457eab","Type":"ContainerDied","Data":"17667fa4b595d4d5f150d54c90b0674248035cd9f0def0674ea10cfc58e20fa2"} Mar 11 13:25:18 crc kubenswrapper[4923]: I0311 13:25:18.936836 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s9ncx" event={"ID":"dbba7a9b-2c9b-4770-ac96-b89390457eab","Type":"ContainerStarted","Data":"3aaec75616cbaf40a150e962bdb9d83abaeb7912184e24154201dd89ca513a24"} Mar 11 13:25:18 crc kubenswrapper[4923]: I0311 13:25:18.940485 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqdq8" event={"ID":"7ce0bb4a-0841-4449-bc02-6c219731272e","Type":"ContainerStarted","Data":"a5489594d4199b5105fab493c51e3da64bd6c7c739f44779631fd6d76a22972a"} Mar 11 13:25:18 crc kubenswrapper[4923]: I0311 13:25:18.940992 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-97n98" podUID="08394455-c98e-4e73-b798-3a9c2ff5e708" containerName="registry-server" containerID="cri-o://6e5fc33d552910006afd99bf5fbbf246fc57409a0d4cd61463d2d4417b5887db" gracePeriod=2 Mar 11 13:25:18 crc kubenswrapper[4923]: I0311 13:25:18.955537 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-s9ncx" podStartSLOduration=2.929836442 podStartE2EDuration="46.955517948s" podCreationTimestamp="2026-03-11 13:24:32 +0000 UTC" firstStartedPulling="2026-03-11 13:24:34.378321672 +0000 UTC m=+122.060223386" lastFinishedPulling="2026-03-11 13:25:18.404003178 +0000 UTC m=+166.085904892" observedRunningTime="2026-03-11 13:25:18.951794326 +0000 UTC m=+166.633696040" watchObservedRunningTime="2026-03-11 13:25:18.955517948 +0000 UTC m=+166.637419672" Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.339282 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-97n98" Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.535321 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08394455-c98e-4e73-b798-3a9c2ff5e708-catalog-content\") pod \"08394455-c98e-4e73-b798-3a9c2ff5e708\" (UID: \"08394455-c98e-4e73-b798-3a9c2ff5e708\") " Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.535452 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08394455-c98e-4e73-b798-3a9c2ff5e708-utilities\") pod \"08394455-c98e-4e73-b798-3a9c2ff5e708\" (UID: \"08394455-c98e-4e73-b798-3a9c2ff5e708\") " Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.535621 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s29bn\" (UniqueName: \"kubernetes.io/projected/08394455-c98e-4e73-b798-3a9c2ff5e708-kube-api-access-s29bn\") pod \"08394455-c98e-4e73-b798-3a9c2ff5e708\" (UID: \"08394455-c98e-4e73-b798-3a9c2ff5e708\") " Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.536390 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08394455-c98e-4e73-b798-3a9c2ff5e708-utilities" (OuterVolumeSpecName: "utilities") pod "08394455-c98e-4e73-b798-3a9c2ff5e708" (UID: "08394455-c98e-4e73-b798-3a9c2ff5e708"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.542657 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08394455-c98e-4e73-b798-3a9c2ff5e708-kube-api-access-s29bn" (OuterVolumeSpecName: "kube-api-access-s29bn") pod "08394455-c98e-4e73-b798-3a9c2ff5e708" (UID: "08394455-c98e-4e73-b798-3a9c2ff5e708"). InnerVolumeSpecName "kube-api-access-s29bn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.636569 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s29bn\" (UniqueName: \"kubernetes.io/projected/08394455-c98e-4e73-b798-3a9c2ff5e708-kube-api-access-s29bn\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.636609 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08394455-c98e-4e73-b798-3a9c2ff5e708-utilities\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.668558 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08394455-c98e-4e73-b798-3a9c2ff5e708-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08394455-c98e-4e73-b798-3a9c2ff5e708" (UID: "08394455-c98e-4e73-b798-3a9c2ff5e708"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.737377 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08394455-c98e-4e73-b798-3a9c2ff5e708-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.760811 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 11 13:25:19 crc kubenswrapper[4923]: E0311 13:25:19.761092 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08394455-c98e-4e73-b798-3a9c2ff5e708" containerName="registry-server" Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.761114 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="08394455-c98e-4e73-b798-3a9c2ff5e708" containerName="registry-server" Mar 11 13:25:19 crc kubenswrapper[4923]: E0311 13:25:19.761137 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08394455-c98e-4e73-b798-3a9c2ff5e708" containerName="extract-content" Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.761146 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="08394455-c98e-4e73-b798-3a9c2ff5e708" containerName="extract-content" Mar 11 13:25:19 crc kubenswrapper[4923]: E0311 13:25:19.761158 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08394455-c98e-4e73-b798-3a9c2ff5e708" containerName="extract-utilities" Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.761167 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="08394455-c98e-4e73-b798-3a9c2ff5e708" containerName="extract-utilities" Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.761316 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="08394455-c98e-4e73-b798-3a9c2ff5e708" containerName="registry-server" Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.761920 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.764264 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.766018 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.768924 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.938921 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/331aaa7c-9dd6-4f2c-b3d8-00febd6448b8-var-lock\") pod \"installer-9-crc\" (UID: \"331aaa7c-9dd6-4f2c-b3d8-00febd6448b8\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.939009 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/331aaa7c-9dd6-4f2c-b3d8-00febd6448b8-kubelet-dir\") pod \"installer-9-crc\" (UID: \"331aaa7c-9dd6-4f2c-b3d8-00febd6448b8\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.939040 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/331aaa7c-9dd6-4f2c-b3d8-00febd6448b8-kube-api-access\") pod \"installer-9-crc\" (UID: \"331aaa7c-9dd6-4f2c-b3d8-00febd6448b8\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.946048 4923 generic.go:334] "Generic (PLEG): container finished" podID="7ce0bb4a-0841-4449-bc02-6c219731272e" containerID="a5489594d4199b5105fab493c51e3da64bd6c7c739f44779631fd6d76a22972a" exitCode=0 Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.946102 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqdq8" event={"ID":"7ce0bb4a-0841-4449-bc02-6c219731272e","Type":"ContainerDied","Data":"a5489594d4199b5105fab493c51e3da64bd6c7c739f44779631fd6d76a22972a"} Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.951596 4923 generic.go:334] "Generic (PLEG): container finished" podID="08394455-c98e-4e73-b798-3a9c2ff5e708" containerID="6e5fc33d552910006afd99bf5fbbf246fc57409a0d4cd61463d2d4417b5887db" exitCode=0 Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.951666 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-97n98" Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.951682 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-97n98" event={"ID":"08394455-c98e-4e73-b798-3a9c2ff5e708","Type":"ContainerDied","Data":"6e5fc33d552910006afd99bf5fbbf246fc57409a0d4cd61463d2d4417b5887db"} Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.952167 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-97n98" event={"ID":"08394455-c98e-4e73-b798-3a9c2ff5e708","Type":"ContainerDied","Data":"298cf8caabfb130792a48b857e4a438ce031ac0223096788367622e427775fbf"} Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.952191 4923 scope.go:117] "RemoveContainer" containerID="6e5fc33d552910006afd99bf5fbbf246fc57409a0d4cd61463d2d4417b5887db" Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.955274 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mcgpl" event={"ID":"9ab29276-3a67-42c9-8308-2ee71a3dd4dc","Type":"ContainerStarted","Data":"97b54193351d8eaf723e7bca0bdbad2eb90e1fb7a9f39032b0d064d296cd1c63"} Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.963058 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdvss" event={"ID":"8bdd5510-3bb1-42f0-8646-c45f11bdeafe","Type":"ContainerStarted","Data":"684d14d54963ecdb7db1a661ffa2bdb391a2b6c263d5a2b213f172d22dd0a68c"} Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.977302 4923 scope.go:117] "RemoveContainer" containerID="f8afffeb3abdc4cefd142f962216f37fbba1758b9046e280c28aef899cb1446f" Mar 11 13:25:19 crc kubenswrapper[4923]: I0311 13:25:19.999153 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rdvss" podStartSLOduration=3.334090673 podStartE2EDuration="47.999132958s" podCreationTimestamp="2026-03-11 13:24:32 +0000 UTC" firstStartedPulling="2026-03-11 13:24:34.369649116 +0000 UTC m=+122.051550830" lastFinishedPulling="2026-03-11 13:25:19.034691401 +0000 UTC m=+166.716593115" observedRunningTime="2026-03-11 13:25:19.996561197 +0000 UTC m=+167.678462931" watchObservedRunningTime="2026-03-11 13:25:19.999132958 +0000 UTC m=+167.681034672" Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.014360 4923 scope.go:117] "RemoveContainer" containerID="c14e06ae906be3c2a21149ec1115cd7fd6b13cb240a32bffd813fae42269b678" Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.034258 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-97n98"] Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.039784 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/331aaa7c-9dd6-4f2c-b3d8-00febd6448b8-kubelet-dir\") pod \"installer-9-crc\" (UID: \"331aaa7c-9dd6-4f2c-b3d8-00febd6448b8\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.039818 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/331aaa7c-9dd6-4f2c-b3d8-00febd6448b8-kube-api-access\") pod \"installer-9-crc\" (UID: \"331aaa7c-9dd6-4f2c-b3d8-00febd6448b8\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.039859 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/331aaa7c-9dd6-4f2c-b3d8-00febd6448b8-kubelet-dir\") pod \"installer-9-crc\" (UID: \"331aaa7c-9dd6-4f2c-b3d8-00febd6448b8\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.039891 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/331aaa7c-9dd6-4f2c-b3d8-00febd6448b8-var-lock\") pod \"installer-9-crc\" (UID: \"331aaa7c-9dd6-4f2c-b3d8-00febd6448b8\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.039963 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/331aaa7c-9dd6-4f2c-b3d8-00febd6448b8-var-lock\") pod \"installer-9-crc\" (UID: \"331aaa7c-9dd6-4f2c-b3d8-00febd6448b8\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.040828 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-97n98"] Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.064263 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/331aaa7c-9dd6-4f2c-b3d8-00febd6448b8-kube-api-access\") pod \"installer-9-crc\" (UID: \"331aaa7c-9dd6-4f2c-b3d8-00febd6448b8\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.097737 4923 scope.go:117] "RemoveContainer" containerID="6e5fc33d552910006afd99bf5fbbf246fc57409a0d4cd61463d2d4417b5887db" Mar 11 13:25:20 crc kubenswrapper[4923]: E0311 13:25:20.098166 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e5fc33d552910006afd99bf5fbbf246fc57409a0d4cd61463d2d4417b5887db\": container with ID starting with 6e5fc33d552910006afd99bf5fbbf246fc57409a0d4cd61463d2d4417b5887db not found: ID does not exist" containerID="6e5fc33d552910006afd99bf5fbbf246fc57409a0d4cd61463d2d4417b5887db" Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.098216 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e5fc33d552910006afd99bf5fbbf246fc57409a0d4cd61463d2d4417b5887db"} err="failed to get container status \"6e5fc33d552910006afd99bf5fbbf246fc57409a0d4cd61463d2d4417b5887db\": rpc error: code = NotFound desc = could not find container \"6e5fc33d552910006afd99bf5fbbf246fc57409a0d4cd61463d2d4417b5887db\": container with ID starting with 6e5fc33d552910006afd99bf5fbbf246fc57409a0d4cd61463d2d4417b5887db not found: ID does not exist" Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.098244 4923 scope.go:117] "RemoveContainer" containerID="f8afffeb3abdc4cefd142f962216f37fbba1758b9046e280c28aef899cb1446f" Mar 11 13:25:20 crc kubenswrapper[4923]: E0311 13:25:20.098687 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8afffeb3abdc4cefd142f962216f37fbba1758b9046e280c28aef899cb1446f\": container with ID starting with f8afffeb3abdc4cefd142f962216f37fbba1758b9046e280c28aef899cb1446f not found: ID does not exist" containerID="f8afffeb3abdc4cefd142f962216f37fbba1758b9046e280c28aef899cb1446f" Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.098734 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8afffeb3abdc4cefd142f962216f37fbba1758b9046e280c28aef899cb1446f"} err="failed to get container status \"f8afffeb3abdc4cefd142f962216f37fbba1758b9046e280c28aef899cb1446f\": rpc error: code = NotFound desc = could not find container \"f8afffeb3abdc4cefd142f962216f37fbba1758b9046e280c28aef899cb1446f\": container with ID starting with f8afffeb3abdc4cefd142f962216f37fbba1758b9046e280c28aef899cb1446f not found: ID does not exist" Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.098765 4923 scope.go:117] "RemoveContainer" containerID="c14e06ae906be3c2a21149ec1115cd7fd6b13cb240a32bffd813fae42269b678" Mar 11 13:25:20 crc kubenswrapper[4923]: E0311 13:25:20.099177 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c14e06ae906be3c2a21149ec1115cd7fd6b13cb240a32bffd813fae42269b678\": container with ID starting with c14e06ae906be3c2a21149ec1115cd7fd6b13cb240a32bffd813fae42269b678 not found: ID does not exist" containerID="c14e06ae906be3c2a21149ec1115cd7fd6b13cb240a32bffd813fae42269b678" Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.099204 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c14e06ae906be3c2a21149ec1115cd7fd6b13cb240a32bffd813fae42269b678"} err="failed to get container status \"c14e06ae906be3c2a21149ec1115cd7fd6b13cb240a32bffd813fae42269b678\": rpc error: code = NotFound desc = could not find container \"c14e06ae906be3c2a21149ec1115cd7fd6b13cb240a32bffd813fae42269b678\": container with ID starting with c14e06ae906be3c2a21149ec1115cd7fd6b13cb240a32bffd813fae42269b678 not found: ID does not exist" Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.111187 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.609830 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.970322 4923 generic.go:334] "Generic (PLEG): container finished" podID="9ab29276-3a67-42c9-8308-2ee71a3dd4dc" containerID="97b54193351d8eaf723e7bca0bdbad2eb90e1fb7a9f39032b0d064d296cd1c63" exitCode=0 Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.970415 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mcgpl" event={"ID":"9ab29276-3a67-42c9-8308-2ee71a3dd4dc","Type":"ContainerDied","Data":"97b54193351d8eaf723e7bca0bdbad2eb90e1fb7a9f39032b0d064d296cd1c63"} Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.973842 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqdq8" event={"ID":"7ce0bb4a-0841-4449-bc02-6c219731272e","Type":"ContainerStarted","Data":"3e492da31d96171c02f8879c5e7423f7a1c454b1fc4d109a1c5db5ab6c4a2c0c"} Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.976601 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"331aaa7c-9dd6-4f2c-b3d8-00febd6448b8","Type":"ContainerStarted","Data":"32f264786f21fe350bcdecb293da2b7dcf1e5cecba0e4ac8023fd116eb0f5a69"} Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.976634 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"331aaa7c-9dd6-4f2c-b3d8-00febd6448b8","Type":"ContainerStarted","Data":"4cef773813c29b7c700536a9913586524b812d852e9058e3c6ff5a9c38e3aed5"} Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.984738 4923 generic.go:334] "Generic (PLEG): container finished" podID="52f59385-23ad-4163-a766-b9d21e57ad6c" containerID="84a62c4f1b65518a0426df3f6f362d9e4629e3340435f67003dc4864fce6b196" exitCode=0 Mar 11 13:25:20 crc kubenswrapper[4923]: I0311 13:25:20.984792 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8f4cb" event={"ID":"52f59385-23ad-4163-a766-b9d21e57ad6c","Type":"ContainerDied","Data":"84a62c4f1b65518a0426df3f6f362d9e4629e3340435f67003dc4864fce6b196"} Mar 11 13:25:21 crc kubenswrapper[4923]: I0311 13:25:21.033826 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08394455-c98e-4e73-b798-3a9c2ff5e708" path="/var/lib/kubelet/pods/08394455-c98e-4e73-b798-3a9c2ff5e708/volumes" Mar 11 13:25:21 crc kubenswrapper[4923]: I0311 13:25:21.034234 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.034219033 podStartE2EDuration="2.034219033s" podCreationTimestamp="2026-03-11 13:25:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:25:21.028677501 +0000 UTC m=+168.710579215" watchObservedRunningTime="2026-03-11 13:25:21.034219033 +0000 UTC m=+168.716120747" Mar 11 13:25:21 crc kubenswrapper[4923]: I0311 13:25:21.065785 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cqdq8" podStartSLOduration=3.072446818 podStartE2EDuration="49.065766298s" podCreationTimestamp="2026-03-11 13:24:32 +0000 UTC" firstStartedPulling="2026-03-11 13:24:34.392641437 +0000 UTC m=+122.074543151" lastFinishedPulling="2026-03-11 13:25:20.385960917 +0000 UTC m=+168.067862631" observedRunningTime="2026-03-11 13:25:21.048522085 +0000 UTC m=+168.730423799" watchObservedRunningTime="2026-03-11 13:25:21.065766298 +0000 UTC m=+168.747668012" Mar 11 13:25:21 crc kubenswrapper[4923]: I0311 13:25:21.992198 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8f4cb" event={"ID":"52f59385-23ad-4163-a766-b9d21e57ad6c","Type":"ContainerStarted","Data":"784602a924d09abfb2e1a681cacf54d66ca459bb02d50b9e2b24ba71177c485a"} Mar 11 13:25:21 crc kubenswrapper[4923]: I0311 13:25:21.995455 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mcgpl" event={"ID":"9ab29276-3a67-42c9-8308-2ee71a3dd4dc","Type":"ContainerStarted","Data":"642415d512db31739f20a8292dd4afebb284c201e1f7627e89a0c88bcea54756"} Mar 11 13:25:21 crc kubenswrapper[4923]: I0311 13:25:21.997205 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jcncp" event={"ID":"ac5c84a5-f129-440d-8423-f8d86b00e3b9","Type":"ContainerStarted","Data":"32c9a726e5c0ba79943abfa9da284d8ef2f3e0beebc78ef5904e35d21a6c483b"} Mar 11 13:25:22 crc kubenswrapper[4923]: I0311 13:25:22.013778 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8f4cb" podStartSLOduration=2.766656783 podStartE2EDuration="50.013753813s" podCreationTimestamp="2026-03-11 13:24:32 +0000 UTC" firstStartedPulling="2026-03-11 13:24:34.362859908 +0000 UTC m=+122.044761632" lastFinishedPulling="2026-03-11 13:25:21.609956948 +0000 UTC m=+169.291858662" observedRunningTime="2026-03-11 13:25:22.012759515 +0000 UTC m=+169.694661229" watchObservedRunningTime="2026-03-11 13:25:22.013753813 +0000 UTC m=+169.695655527" Mar 11 13:25:22 crc kubenswrapper[4923]: I0311 13:25:22.045127 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mcgpl" podStartSLOduration=4.075578167 podStartE2EDuration="48.045109964s" podCreationTimestamp="2026-03-11 13:24:34 +0000 UTC" firstStartedPulling="2026-03-11 13:24:37.518488273 +0000 UTC m=+125.200389987" lastFinishedPulling="2026-03-11 13:25:21.48802007 +0000 UTC m=+169.169921784" observedRunningTime="2026-03-11 13:25:22.042578284 +0000 UTC m=+169.724479998" watchObservedRunningTime="2026-03-11 13:25:22.045109964 +0000 UTC m=+169.727011678" Mar 11 13:25:22 crc kubenswrapper[4923]: I0311 13:25:22.745789 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8f4cb" Mar 11 13:25:22 crc kubenswrapper[4923]: I0311 13:25:22.746160 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8f4cb" Mar 11 13:25:22 crc kubenswrapper[4923]: I0311 13:25:22.889726 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cqdq8" Mar 11 13:25:22 crc kubenswrapper[4923]: I0311 13:25:22.889780 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cqdq8" Mar 11 13:25:22 crc kubenswrapper[4923]: I0311 13:25:22.928355 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cqdq8" Mar 11 13:25:23 crc kubenswrapper[4923]: I0311 13:25:23.005180 4923 generic.go:334] "Generic (PLEG): container finished" podID="ac5c84a5-f129-440d-8423-f8d86b00e3b9" containerID="32c9a726e5c0ba79943abfa9da284d8ef2f3e0beebc78ef5904e35d21a6c483b" exitCode=0 Mar 11 13:25:23 crc kubenswrapper[4923]: I0311 13:25:23.005256 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jcncp" event={"ID":"ac5c84a5-f129-440d-8423-f8d86b00e3b9","Type":"ContainerDied","Data":"32c9a726e5c0ba79943abfa9da284d8ef2f3e0beebc78ef5904e35d21a6c483b"} Mar 11 13:25:23 crc kubenswrapper[4923]: I0311 13:25:23.229160 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rdvss" Mar 11 13:25:23 crc kubenswrapper[4923]: I0311 13:25:23.229203 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rdvss" Mar 11 13:25:23 crc kubenswrapper[4923]: I0311 13:25:23.280506 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rdvss" Mar 11 13:25:23 crc kubenswrapper[4923]: I0311 13:25:23.452700 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-s9ncx" Mar 11 13:25:23 crc kubenswrapper[4923]: I0311 13:25:23.452755 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-s9ncx" Mar 11 13:25:23 crc kubenswrapper[4923]: I0311 13:25:23.495529 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-s9ncx" Mar 11 13:25:23 crc kubenswrapper[4923]: I0311 13:25:23.792315 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-8f4cb" podUID="52f59385-23ad-4163-a766-b9d21e57ad6c" containerName="registry-server" probeResult="failure" output=< Mar 11 13:25:23 crc kubenswrapper[4923]: timeout: failed to connect service ":50051" within 1s Mar 11 13:25:23 crc kubenswrapper[4923]: > Mar 11 13:25:24 crc kubenswrapper[4923]: I0311 13:25:24.014925 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jcncp" event={"ID":"ac5c84a5-f129-440d-8423-f8d86b00e3b9","Type":"ContainerStarted","Data":"97442e068883c28e9c46d4062a84e4429f6597de785028bde4b003dfc13d63d6"} Mar 11 13:25:24 crc kubenswrapper[4923]: I0311 13:25:24.070193 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-s9ncx" Mar 11 13:25:24 crc kubenswrapper[4923]: I0311 13:25:24.072073 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rdvss" Mar 11 13:25:24 crc kubenswrapper[4923]: I0311 13:25:24.099334 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jcncp" podStartSLOduration=2.02018945 podStartE2EDuration="50.099314976s" podCreationTimestamp="2026-03-11 13:24:34 +0000 UTC" firstStartedPulling="2026-03-11 13:24:35.444849326 +0000 UTC m=+123.126751040" lastFinishedPulling="2026-03-11 13:25:23.523974852 +0000 UTC m=+171.205876566" observedRunningTime="2026-03-11 13:25:24.04225981 +0000 UTC m=+171.724161524" watchObservedRunningTime="2026-03-11 13:25:24.099314976 +0000 UTC m=+171.781216690" Mar 11 13:25:24 crc kubenswrapper[4923]: I0311 13:25:24.901015 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jcncp" Mar 11 13:25:24 crc kubenswrapper[4923]: I0311 13:25:24.901969 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jcncp" Mar 11 13:25:25 crc kubenswrapper[4923]: I0311 13:25:25.269886 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s9ncx"] Mar 11 13:25:25 crc kubenswrapper[4923]: I0311 13:25:25.278766 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mcgpl" Mar 11 13:25:25 crc kubenswrapper[4923]: I0311 13:25:25.278847 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mcgpl" Mar 11 13:25:25 crc kubenswrapper[4923]: I0311 13:25:25.338475 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mcgpl" Mar 11 13:25:25 crc kubenswrapper[4923]: I0311 13:25:25.866919 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rdvss"] Mar 11 13:25:25 crc kubenswrapper[4923]: I0311 13:25:25.949680 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-jcncp" podUID="ac5c84a5-f129-440d-8423-f8d86b00e3b9" containerName="registry-server" probeResult="failure" output=< Mar 11 13:25:25 crc kubenswrapper[4923]: timeout: failed to connect service ":50051" within 1s Mar 11 13:25:25 crc kubenswrapper[4923]: > Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.027638 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rdvss" podUID="8bdd5510-3bb1-42f0-8646-c45f11bdeafe" containerName="registry-server" containerID="cri-o://684d14d54963ecdb7db1a661ffa2bdb391a2b6c263d5a2b213f172d22dd0a68c" gracePeriod=2 Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.027848 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-s9ncx" podUID="dbba7a9b-2c9b-4770-ac96-b89390457eab" containerName="registry-server" containerID="cri-o://3aaec75616cbaf40a150e962bdb9d83abaeb7912184e24154201dd89ca513a24" gracePeriod=2 Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.071432 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mcgpl" Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.560710 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s9ncx" Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.604155 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rdvss" Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.656886 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bdd5510-3bb1-42f0-8646-c45f11bdeafe-utilities\") pod \"8bdd5510-3bb1-42f0-8646-c45f11bdeafe\" (UID: \"8bdd5510-3bb1-42f0-8646-c45f11bdeafe\") " Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.657010 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbba7a9b-2c9b-4770-ac96-b89390457eab-utilities\") pod \"dbba7a9b-2c9b-4770-ac96-b89390457eab\" (UID: \"dbba7a9b-2c9b-4770-ac96-b89390457eab\") " Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.657063 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbba7a9b-2c9b-4770-ac96-b89390457eab-catalog-content\") pod \"dbba7a9b-2c9b-4770-ac96-b89390457eab\" (UID: \"dbba7a9b-2c9b-4770-ac96-b89390457eab\") " Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.657126 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w48wt\" (UniqueName: \"kubernetes.io/projected/dbba7a9b-2c9b-4770-ac96-b89390457eab-kube-api-access-w48wt\") pod \"dbba7a9b-2c9b-4770-ac96-b89390457eab\" (UID: \"dbba7a9b-2c9b-4770-ac96-b89390457eab\") " Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.657171 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drk5r\" (UniqueName: \"kubernetes.io/projected/8bdd5510-3bb1-42f0-8646-c45f11bdeafe-kube-api-access-drk5r\") pod \"8bdd5510-3bb1-42f0-8646-c45f11bdeafe\" (UID: \"8bdd5510-3bb1-42f0-8646-c45f11bdeafe\") " Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.657238 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bdd5510-3bb1-42f0-8646-c45f11bdeafe-catalog-content\") pod \"8bdd5510-3bb1-42f0-8646-c45f11bdeafe\" (UID: \"8bdd5510-3bb1-42f0-8646-c45f11bdeafe\") " Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.658824 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbba7a9b-2c9b-4770-ac96-b89390457eab-utilities" (OuterVolumeSpecName: "utilities") pod "dbba7a9b-2c9b-4770-ac96-b89390457eab" (UID: "dbba7a9b-2c9b-4770-ac96-b89390457eab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.659690 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bdd5510-3bb1-42f0-8646-c45f11bdeafe-utilities" (OuterVolumeSpecName: "utilities") pod "8bdd5510-3bb1-42f0-8646-c45f11bdeafe" (UID: "8bdd5510-3bb1-42f0-8646-c45f11bdeafe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.665503 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbba7a9b-2c9b-4770-ac96-b89390457eab-kube-api-access-w48wt" (OuterVolumeSpecName: "kube-api-access-w48wt") pod "dbba7a9b-2c9b-4770-ac96-b89390457eab" (UID: "dbba7a9b-2c9b-4770-ac96-b89390457eab"). InnerVolumeSpecName "kube-api-access-w48wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.665644 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bdd5510-3bb1-42f0-8646-c45f11bdeafe-kube-api-access-drk5r" (OuterVolumeSpecName: "kube-api-access-drk5r") pod "8bdd5510-3bb1-42f0-8646-c45f11bdeafe" (UID: "8bdd5510-3bb1-42f0-8646-c45f11bdeafe"). InnerVolumeSpecName "kube-api-access-drk5r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.728945 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbba7a9b-2c9b-4770-ac96-b89390457eab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dbba7a9b-2c9b-4770-ac96-b89390457eab" (UID: "dbba7a9b-2c9b-4770-ac96-b89390457eab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.759951 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w48wt\" (UniqueName: \"kubernetes.io/projected/dbba7a9b-2c9b-4770-ac96-b89390457eab-kube-api-access-w48wt\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.760012 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drk5r\" (UniqueName: \"kubernetes.io/projected/8bdd5510-3bb1-42f0-8646-c45f11bdeafe-kube-api-access-drk5r\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.760026 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8bdd5510-3bb1-42f0-8646-c45f11bdeafe-utilities\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.760054 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbba7a9b-2c9b-4770-ac96-b89390457eab-utilities\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.760070 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbba7a9b-2c9b-4770-ac96-b89390457eab-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.809409 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bdd5510-3bb1-42f0-8646-c45f11bdeafe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8bdd5510-3bb1-42f0-8646-c45f11bdeafe" (UID: "8bdd5510-3bb1-42f0-8646-c45f11bdeafe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:25:26 crc kubenswrapper[4923]: I0311 13:25:26.861691 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8bdd5510-3bb1-42f0-8646-c45f11bdeafe-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.034288 4923 generic.go:334] "Generic (PLEG): container finished" podID="8bdd5510-3bb1-42f0-8646-c45f11bdeafe" containerID="684d14d54963ecdb7db1a661ffa2bdb391a2b6c263d5a2b213f172d22dd0a68c" exitCode=0 Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.034410 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdvss" event={"ID":"8bdd5510-3bb1-42f0-8646-c45f11bdeafe","Type":"ContainerDied","Data":"684d14d54963ecdb7db1a661ffa2bdb391a2b6c263d5a2b213f172d22dd0a68c"} Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.034857 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rdvss" event={"ID":"8bdd5510-3bb1-42f0-8646-c45f11bdeafe","Type":"ContainerDied","Data":"28506726bb33e4937d4c58ccc316a19ff456adddcd5a84b46308d237de1b7cdf"} Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.034905 4923 scope.go:117] "RemoveContainer" containerID="684d14d54963ecdb7db1a661ffa2bdb391a2b6c263d5a2b213f172d22dd0a68c" Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.034469 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rdvss" Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.043949 4923 generic.go:334] "Generic (PLEG): container finished" podID="dbba7a9b-2c9b-4770-ac96-b89390457eab" containerID="3aaec75616cbaf40a150e962bdb9d83abaeb7912184e24154201dd89ca513a24" exitCode=0 Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.044166 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s9ncx" event={"ID":"dbba7a9b-2c9b-4770-ac96-b89390457eab","Type":"ContainerDied","Data":"3aaec75616cbaf40a150e962bdb9d83abaeb7912184e24154201dd89ca513a24"} Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.044249 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s9ncx" event={"ID":"dbba7a9b-2c9b-4770-ac96-b89390457eab","Type":"ContainerDied","Data":"184cf4856a6dd957f402dab4aa900b8d61954d494dad4611ca54a419a5cf33f1"} Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.044443 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s9ncx" Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.079978 4923 scope.go:117] "RemoveContainer" containerID="436dc8cf923342891f9bf2a15c47e721e0f5b23f0b3e92e7bea7239d8a46ff47" Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.085714 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s9ncx"] Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.089153 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-s9ncx"] Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.097181 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rdvss"] Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.099326 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rdvss"] Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.100826 4923 scope.go:117] "RemoveContainer" containerID="e73ae67b9317111c1da0092fec6fb392e4d3eb633bc285917b7dd84798ca2a57" Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.114962 4923 scope.go:117] "RemoveContainer" containerID="684d14d54963ecdb7db1a661ffa2bdb391a2b6c263d5a2b213f172d22dd0a68c" Mar 11 13:25:27 crc kubenswrapper[4923]: E0311 13:25:27.115372 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"684d14d54963ecdb7db1a661ffa2bdb391a2b6c263d5a2b213f172d22dd0a68c\": container with ID starting with 684d14d54963ecdb7db1a661ffa2bdb391a2b6c263d5a2b213f172d22dd0a68c not found: ID does not exist" containerID="684d14d54963ecdb7db1a661ffa2bdb391a2b6c263d5a2b213f172d22dd0a68c" Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.115443 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"684d14d54963ecdb7db1a661ffa2bdb391a2b6c263d5a2b213f172d22dd0a68c"} err="failed to get container status \"684d14d54963ecdb7db1a661ffa2bdb391a2b6c263d5a2b213f172d22dd0a68c\": rpc error: code = NotFound desc = could not find container \"684d14d54963ecdb7db1a661ffa2bdb391a2b6c263d5a2b213f172d22dd0a68c\": container with ID starting with 684d14d54963ecdb7db1a661ffa2bdb391a2b6c263d5a2b213f172d22dd0a68c not found: ID does not exist" Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.115524 4923 scope.go:117] "RemoveContainer" containerID="436dc8cf923342891f9bf2a15c47e721e0f5b23f0b3e92e7bea7239d8a46ff47" Mar 11 13:25:27 crc kubenswrapper[4923]: E0311 13:25:27.122505 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"436dc8cf923342891f9bf2a15c47e721e0f5b23f0b3e92e7bea7239d8a46ff47\": container with ID starting with 436dc8cf923342891f9bf2a15c47e721e0f5b23f0b3e92e7bea7239d8a46ff47 not found: ID does not exist" containerID="436dc8cf923342891f9bf2a15c47e721e0f5b23f0b3e92e7bea7239d8a46ff47" Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.122526 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"436dc8cf923342891f9bf2a15c47e721e0f5b23f0b3e92e7bea7239d8a46ff47"} err="failed to get container status \"436dc8cf923342891f9bf2a15c47e721e0f5b23f0b3e92e7bea7239d8a46ff47\": rpc error: code = NotFound desc = could not find container \"436dc8cf923342891f9bf2a15c47e721e0f5b23f0b3e92e7bea7239d8a46ff47\": container with ID starting with 436dc8cf923342891f9bf2a15c47e721e0f5b23f0b3e92e7bea7239d8a46ff47 not found: ID does not exist" Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.122541 4923 scope.go:117] "RemoveContainer" containerID="e73ae67b9317111c1da0092fec6fb392e4d3eb633bc285917b7dd84798ca2a57" Mar 11 13:25:27 crc kubenswrapper[4923]: E0311 13:25:27.122765 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e73ae67b9317111c1da0092fec6fb392e4d3eb633bc285917b7dd84798ca2a57\": container with ID starting with e73ae67b9317111c1da0092fec6fb392e4d3eb633bc285917b7dd84798ca2a57 not found: ID does not exist" containerID="e73ae67b9317111c1da0092fec6fb392e4d3eb633bc285917b7dd84798ca2a57" Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.122786 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e73ae67b9317111c1da0092fec6fb392e4d3eb633bc285917b7dd84798ca2a57"} err="failed to get container status \"e73ae67b9317111c1da0092fec6fb392e4d3eb633bc285917b7dd84798ca2a57\": rpc error: code = NotFound desc = could not find container \"e73ae67b9317111c1da0092fec6fb392e4d3eb633bc285917b7dd84798ca2a57\": container with ID starting with e73ae67b9317111c1da0092fec6fb392e4d3eb633bc285917b7dd84798ca2a57 not found: ID does not exist" Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.122799 4923 scope.go:117] "RemoveContainer" containerID="3aaec75616cbaf40a150e962bdb9d83abaeb7912184e24154201dd89ca513a24" Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.137258 4923 scope.go:117] "RemoveContainer" containerID="17667fa4b595d4d5f150d54c90b0674248035cd9f0def0674ea10cfc58e20fa2" Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.161777 4923 scope.go:117] "RemoveContainer" containerID="cf4922608027344a703c1ef9f815054a3bdba81a21392a04abf1e2d76d3d244b" Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.178516 4923 scope.go:117] "RemoveContainer" containerID="3aaec75616cbaf40a150e962bdb9d83abaeb7912184e24154201dd89ca513a24" Mar 11 13:25:27 crc kubenswrapper[4923]: E0311 13:25:27.179157 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3aaec75616cbaf40a150e962bdb9d83abaeb7912184e24154201dd89ca513a24\": container with ID starting with 3aaec75616cbaf40a150e962bdb9d83abaeb7912184e24154201dd89ca513a24 not found: ID does not exist" containerID="3aaec75616cbaf40a150e962bdb9d83abaeb7912184e24154201dd89ca513a24" Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.179266 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3aaec75616cbaf40a150e962bdb9d83abaeb7912184e24154201dd89ca513a24"} err="failed to get container status \"3aaec75616cbaf40a150e962bdb9d83abaeb7912184e24154201dd89ca513a24\": rpc error: code = NotFound desc = could not find container \"3aaec75616cbaf40a150e962bdb9d83abaeb7912184e24154201dd89ca513a24\": container with ID starting with 3aaec75616cbaf40a150e962bdb9d83abaeb7912184e24154201dd89ca513a24 not found: ID does not exist" Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.179386 4923 scope.go:117] "RemoveContainer" containerID="17667fa4b595d4d5f150d54c90b0674248035cd9f0def0674ea10cfc58e20fa2" Mar 11 13:25:27 crc kubenswrapper[4923]: E0311 13:25:27.179888 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17667fa4b595d4d5f150d54c90b0674248035cd9f0def0674ea10cfc58e20fa2\": container with ID starting with 17667fa4b595d4d5f150d54c90b0674248035cd9f0def0674ea10cfc58e20fa2 not found: ID does not exist" containerID="17667fa4b595d4d5f150d54c90b0674248035cd9f0def0674ea10cfc58e20fa2" Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.179930 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17667fa4b595d4d5f150d54c90b0674248035cd9f0def0674ea10cfc58e20fa2"} err="failed to get container status \"17667fa4b595d4d5f150d54c90b0674248035cd9f0def0674ea10cfc58e20fa2\": rpc error: code = NotFound desc = could not find container \"17667fa4b595d4d5f150d54c90b0674248035cd9f0def0674ea10cfc58e20fa2\": container with ID starting with 17667fa4b595d4d5f150d54c90b0674248035cd9f0def0674ea10cfc58e20fa2 not found: ID does not exist" Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.179957 4923 scope.go:117] "RemoveContainer" containerID="cf4922608027344a703c1ef9f815054a3bdba81a21392a04abf1e2d76d3d244b" Mar 11 13:25:27 crc kubenswrapper[4923]: E0311 13:25:27.180220 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf4922608027344a703c1ef9f815054a3bdba81a21392a04abf1e2d76d3d244b\": container with ID starting with cf4922608027344a703c1ef9f815054a3bdba81a21392a04abf1e2d76d3d244b not found: ID does not exist" containerID="cf4922608027344a703c1ef9f815054a3bdba81a21392a04abf1e2d76d3d244b" Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.180243 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf4922608027344a703c1ef9f815054a3bdba81a21392a04abf1e2d76d3d244b"} err="failed to get container status \"cf4922608027344a703c1ef9f815054a3bdba81a21392a04abf1e2d76d3d244b\": rpc error: code = NotFound desc = could not find container \"cf4922608027344a703c1ef9f815054a3bdba81a21392a04abf1e2d76d3d244b\": container with ID starting with cf4922608027344a703c1ef9f815054a3bdba81a21392a04abf1e2d76d3d244b not found: ID does not exist" Mar 11 13:25:27 crc kubenswrapper[4923]: I0311 13:25:27.678449 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mcgpl"] Mar 11 13:25:28 crc kubenswrapper[4923]: I0311 13:25:28.057253 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mcgpl" podUID="9ab29276-3a67-42c9-8308-2ee71a3dd4dc" containerName="registry-server" containerID="cri-o://642415d512db31739f20a8292dd4afebb284c201e1f7627e89a0c88bcea54756" gracePeriod=2 Mar 11 13:25:28 crc kubenswrapper[4923]: I0311 13:25:28.609791 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mcgpl" Mar 11 13:25:28 crc kubenswrapper[4923]: I0311 13:25:28.695019 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnsfj\" (UniqueName: \"kubernetes.io/projected/9ab29276-3a67-42c9-8308-2ee71a3dd4dc-kube-api-access-hnsfj\") pod \"9ab29276-3a67-42c9-8308-2ee71a3dd4dc\" (UID: \"9ab29276-3a67-42c9-8308-2ee71a3dd4dc\") " Mar 11 13:25:28 crc kubenswrapper[4923]: I0311 13:25:28.695072 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ab29276-3a67-42c9-8308-2ee71a3dd4dc-catalog-content\") pod \"9ab29276-3a67-42c9-8308-2ee71a3dd4dc\" (UID: \"9ab29276-3a67-42c9-8308-2ee71a3dd4dc\") " Mar 11 13:25:28 crc kubenswrapper[4923]: I0311 13:25:28.695152 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ab29276-3a67-42c9-8308-2ee71a3dd4dc-utilities\") pod \"9ab29276-3a67-42c9-8308-2ee71a3dd4dc\" (UID: \"9ab29276-3a67-42c9-8308-2ee71a3dd4dc\") " Mar 11 13:25:28 crc kubenswrapper[4923]: I0311 13:25:28.701003 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ab29276-3a67-42c9-8308-2ee71a3dd4dc-utilities" (OuterVolumeSpecName: "utilities") pod "9ab29276-3a67-42c9-8308-2ee71a3dd4dc" (UID: "9ab29276-3a67-42c9-8308-2ee71a3dd4dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:25:28 crc kubenswrapper[4923]: I0311 13:25:28.705051 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ab29276-3a67-42c9-8308-2ee71a3dd4dc-kube-api-access-hnsfj" (OuterVolumeSpecName: "kube-api-access-hnsfj") pod "9ab29276-3a67-42c9-8308-2ee71a3dd4dc" (UID: "9ab29276-3a67-42c9-8308-2ee71a3dd4dc"). InnerVolumeSpecName "kube-api-access-hnsfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:25:28 crc kubenswrapper[4923]: I0311 13:25:28.723709 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ab29276-3a67-42c9-8308-2ee71a3dd4dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9ab29276-3a67-42c9-8308-2ee71a3dd4dc" (UID: "9ab29276-3a67-42c9-8308-2ee71a3dd4dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:25:28 crc kubenswrapper[4923]: I0311 13:25:28.796252 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnsfj\" (UniqueName: \"kubernetes.io/projected/9ab29276-3a67-42c9-8308-2ee71a3dd4dc-kube-api-access-hnsfj\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:28 crc kubenswrapper[4923]: I0311 13:25:28.796305 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ab29276-3a67-42c9-8308-2ee71a3dd4dc-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:28 crc kubenswrapper[4923]: I0311 13:25:28.796324 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ab29276-3a67-42c9-8308-2ee71a3dd4dc-utilities\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:29 crc kubenswrapper[4923]: I0311 13:25:29.031096 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bdd5510-3bb1-42f0-8646-c45f11bdeafe" path="/var/lib/kubelet/pods/8bdd5510-3bb1-42f0-8646-c45f11bdeafe/volumes" Mar 11 13:25:29 crc kubenswrapper[4923]: I0311 13:25:29.031691 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbba7a9b-2c9b-4770-ac96-b89390457eab" path="/var/lib/kubelet/pods/dbba7a9b-2c9b-4770-ac96-b89390457eab/volumes" Mar 11 13:25:29 crc kubenswrapper[4923]: I0311 13:25:29.067588 4923 generic.go:334] "Generic (PLEG): container finished" podID="9ab29276-3a67-42c9-8308-2ee71a3dd4dc" containerID="642415d512db31739f20a8292dd4afebb284c201e1f7627e89a0c88bcea54756" exitCode=0 Mar 11 13:25:29 crc kubenswrapper[4923]: I0311 13:25:29.067653 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mcgpl" event={"ID":"9ab29276-3a67-42c9-8308-2ee71a3dd4dc","Type":"ContainerDied","Data":"642415d512db31739f20a8292dd4afebb284c201e1f7627e89a0c88bcea54756"} Mar 11 13:25:29 crc kubenswrapper[4923]: I0311 13:25:29.067696 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mcgpl" event={"ID":"9ab29276-3a67-42c9-8308-2ee71a3dd4dc","Type":"ContainerDied","Data":"4267893edf7067807fb0c8157225ea1ab1879a1bc50e3c7f871d804888d2565b"} Mar 11 13:25:29 crc kubenswrapper[4923]: I0311 13:25:29.067703 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mcgpl" Mar 11 13:25:29 crc kubenswrapper[4923]: I0311 13:25:29.067727 4923 scope.go:117] "RemoveContainer" containerID="642415d512db31739f20a8292dd4afebb284c201e1f7627e89a0c88bcea54756" Mar 11 13:25:29 crc kubenswrapper[4923]: I0311 13:25:29.093660 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mcgpl"] Mar 11 13:25:29 crc kubenswrapper[4923]: I0311 13:25:29.094068 4923 scope.go:117] "RemoveContainer" containerID="97b54193351d8eaf723e7bca0bdbad2eb90e1fb7a9f39032b0d064d296cd1c63" Mar 11 13:25:29 crc kubenswrapper[4923]: I0311 13:25:29.096871 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mcgpl"] Mar 11 13:25:29 crc kubenswrapper[4923]: I0311 13:25:29.113466 4923 scope.go:117] "RemoveContainer" containerID="c8462f8718db09f8a61636a049de2e143ca3bf015bf2d3f505f39ac804f9d295" Mar 11 13:25:29 crc kubenswrapper[4923]: I0311 13:25:29.139762 4923 scope.go:117] "RemoveContainer" containerID="642415d512db31739f20a8292dd4afebb284c201e1f7627e89a0c88bcea54756" Mar 11 13:25:29 crc kubenswrapper[4923]: E0311 13:25:29.140173 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"642415d512db31739f20a8292dd4afebb284c201e1f7627e89a0c88bcea54756\": container with ID starting with 642415d512db31739f20a8292dd4afebb284c201e1f7627e89a0c88bcea54756 not found: ID does not exist" containerID="642415d512db31739f20a8292dd4afebb284c201e1f7627e89a0c88bcea54756" Mar 11 13:25:29 crc kubenswrapper[4923]: I0311 13:25:29.140209 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"642415d512db31739f20a8292dd4afebb284c201e1f7627e89a0c88bcea54756"} err="failed to get container status \"642415d512db31739f20a8292dd4afebb284c201e1f7627e89a0c88bcea54756\": rpc error: code = NotFound desc = could not find container \"642415d512db31739f20a8292dd4afebb284c201e1f7627e89a0c88bcea54756\": container with ID starting with 642415d512db31739f20a8292dd4afebb284c201e1f7627e89a0c88bcea54756 not found: ID does not exist" Mar 11 13:25:29 crc kubenswrapper[4923]: I0311 13:25:29.140237 4923 scope.go:117] "RemoveContainer" containerID="97b54193351d8eaf723e7bca0bdbad2eb90e1fb7a9f39032b0d064d296cd1c63" Mar 11 13:25:29 crc kubenswrapper[4923]: E0311 13:25:29.140574 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97b54193351d8eaf723e7bca0bdbad2eb90e1fb7a9f39032b0d064d296cd1c63\": container with ID starting with 97b54193351d8eaf723e7bca0bdbad2eb90e1fb7a9f39032b0d064d296cd1c63 not found: ID does not exist" containerID="97b54193351d8eaf723e7bca0bdbad2eb90e1fb7a9f39032b0d064d296cd1c63" Mar 11 13:25:29 crc kubenswrapper[4923]: I0311 13:25:29.140599 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97b54193351d8eaf723e7bca0bdbad2eb90e1fb7a9f39032b0d064d296cd1c63"} err="failed to get container status \"97b54193351d8eaf723e7bca0bdbad2eb90e1fb7a9f39032b0d064d296cd1c63\": rpc error: code = NotFound desc = could not find container \"97b54193351d8eaf723e7bca0bdbad2eb90e1fb7a9f39032b0d064d296cd1c63\": container with ID starting with 97b54193351d8eaf723e7bca0bdbad2eb90e1fb7a9f39032b0d064d296cd1c63 not found: ID does not exist" Mar 11 13:25:29 crc kubenswrapper[4923]: I0311 13:25:29.140618 4923 scope.go:117] "RemoveContainer" containerID="c8462f8718db09f8a61636a049de2e143ca3bf015bf2d3f505f39ac804f9d295" Mar 11 13:25:29 crc kubenswrapper[4923]: E0311 13:25:29.140898 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8462f8718db09f8a61636a049de2e143ca3bf015bf2d3f505f39ac804f9d295\": container with ID starting with c8462f8718db09f8a61636a049de2e143ca3bf015bf2d3f505f39ac804f9d295 not found: ID does not exist" containerID="c8462f8718db09f8a61636a049de2e143ca3bf015bf2d3f505f39ac804f9d295" Mar 11 13:25:29 crc kubenswrapper[4923]: I0311 13:25:29.140924 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8462f8718db09f8a61636a049de2e143ca3bf015bf2d3f505f39ac804f9d295"} err="failed to get container status \"c8462f8718db09f8a61636a049de2e143ca3bf015bf2d3f505f39ac804f9d295\": rpc error: code = NotFound desc = could not find container \"c8462f8718db09f8a61636a049de2e143ca3bf015bf2d3f505f39ac804f9d295\": container with ID starting with c8462f8718db09f8a61636a049de2e143ca3bf015bf2d3f505f39ac804f9d295 not found: ID does not exist" Mar 11 13:25:31 crc kubenswrapper[4923]: I0311 13:25:31.033376 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ab29276-3a67-42c9-8308-2ee71a3dd4dc" path="/var/lib/kubelet/pods/9ab29276-3a67-42c9-8308-2ee71a3dd4dc/volumes" Mar 11 13:25:31 crc kubenswrapper[4923]: I0311 13:25:31.046709 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.081048 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7f5b598bff-5r5gv"] Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.081719 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" podUID="1100409d-07f5-424b-8b9d-f8e8bd066a5c" containerName="controller-manager" containerID="cri-o://04ee7610f6b150a515d1dcae9289d9fcb3dcc44708dffead179693c530f02725" gracePeriod=30 Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.108160 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=1.108138275 podStartE2EDuration="1.108138275s" podCreationTimestamp="2026-03-11 13:25:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:25:32.10506823 +0000 UTC m=+179.786969954" watchObservedRunningTime="2026-03-11 13:25:32.108138275 +0000 UTC m=+179.790040009" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.108548 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk"] Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.108870 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" podUID="13ae0f7e-b016-4044-b6db-4210a79854ab" containerName="route-controller-manager" containerID="cri-o://b7ad9e1425218d639b26e21c5bcb52eab894a864e3e4781633b36fb82ba8826f" gracePeriod=30 Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.658425 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.667799 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.752717 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1100409d-07f5-424b-8b9d-f8e8bd066a5c-client-ca\") pod \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\" (UID: \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\") " Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.753494 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1100409d-07f5-424b-8b9d-f8e8bd066a5c-client-ca" (OuterVolumeSpecName: "client-ca") pod "1100409d-07f5-424b-8b9d-f8e8bd066a5c" (UID: "1100409d-07f5-424b-8b9d-f8e8bd066a5c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.753611 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13ae0f7e-b016-4044-b6db-4210a79854ab-serving-cert\") pod \"13ae0f7e-b016-4044-b6db-4210a79854ab\" (UID: \"13ae0f7e-b016-4044-b6db-4210a79854ab\") " Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.754614 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpw2r\" (UniqueName: \"kubernetes.io/projected/1100409d-07f5-424b-8b9d-f8e8bd066a5c-kube-api-access-zpw2r\") pod \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\" (UID: \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\") " Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.754684 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13ae0f7e-b016-4044-b6db-4210a79854ab-client-ca\") pod \"13ae0f7e-b016-4044-b6db-4210a79854ab\" (UID: \"13ae0f7e-b016-4044-b6db-4210a79854ab\") " Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.754711 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1100409d-07f5-424b-8b9d-f8e8bd066a5c-config\") pod \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\" (UID: \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\") " Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.754739 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2zwt\" (UniqueName: \"kubernetes.io/projected/13ae0f7e-b016-4044-b6db-4210a79854ab-kube-api-access-b2zwt\") pod \"13ae0f7e-b016-4044-b6db-4210a79854ab\" (UID: \"13ae0f7e-b016-4044-b6db-4210a79854ab\") " Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.754773 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13ae0f7e-b016-4044-b6db-4210a79854ab-config\") pod \"13ae0f7e-b016-4044-b6db-4210a79854ab\" (UID: \"13ae0f7e-b016-4044-b6db-4210a79854ab\") " Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.754797 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1100409d-07f5-424b-8b9d-f8e8bd066a5c-proxy-ca-bundles\") pod \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\" (UID: \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\") " Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.754856 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1100409d-07f5-424b-8b9d-f8e8bd066a5c-serving-cert\") pod \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\" (UID: \"1100409d-07f5-424b-8b9d-f8e8bd066a5c\") " Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.755333 4923 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1100409d-07f5-424b-8b9d-f8e8bd066a5c-client-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.755945 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13ae0f7e-b016-4044-b6db-4210a79854ab-client-ca" (OuterVolumeSpecName: "client-ca") pod "13ae0f7e-b016-4044-b6db-4210a79854ab" (UID: "13ae0f7e-b016-4044-b6db-4210a79854ab"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.756260 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1100409d-07f5-424b-8b9d-f8e8bd066a5c-config" (OuterVolumeSpecName: "config") pod "1100409d-07f5-424b-8b9d-f8e8bd066a5c" (UID: "1100409d-07f5-424b-8b9d-f8e8bd066a5c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.756517 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/13ae0f7e-b016-4044-b6db-4210a79854ab-config" (OuterVolumeSpecName: "config") pod "13ae0f7e-b016-4044-b6db-4210a79854ab" (UID: "13ae0f7e-b016-4044-b6db-4210a79854ab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.756899 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1100409d-07f5-424b-8b9d-f8e8bd066a5c-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1100409d-07f5-424b-8b9d-f8e8bd066a5c" (UID: "1100409d-07f5-424b-8b9d-f8e8bd066a5c"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.759111 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13ae0f7e-b016-4044-b6db-4210a79854ab-kube-api-access-b2zwt" (OuterVolumeSpecName: "kube-api-access-b2zwt") pod "13ae0f7e-b016-4044-b6db-4210a79854ab" (UID: "13ae0f7e-b016-4044-b6db-4210a79854ab"). InnerVolumeSpecName "kube-api-access-b2zwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.759205 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13ae0f7e-b016-4044-b6db-4210a79854ab-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "13ae0f7e-b016-4044-b6db-4210a79854ab" (UID: "13ae0f7e-b016-4044-b6db-4210a79854ab"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.759361 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1100409d-07f5-424b-8b9d-f8e8bd066a5c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1100409d-07f5-424b-8b9d-f8e8bd066a5c" (UID: "1100409d-07f5-424b-8b9d-f8e8bd066a5c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.759531 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1100409d-07f5-424b-8b9d-f8e8bd066a5c-kube-api-access-zpw2r" (OuterVolumeSpecName: "kube-api-access-zpw2r") pod "1100409d-07f5-424b-8b9d-f8e8bd066a5c" (UID: "1100409d-07f5-424b-8b9d-f8e8bd066a5c"). InnerVolumeSpecName "kube-api-access-zpw2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.802723 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8f4cb" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.845495 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8f4cb" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.856515 4923 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/13ae0f7e-b016-4044-b6db-4210a79854ab-client-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.856547 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1100409d-07f5-424b-8b9d-f8e8bd066a5c-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.856559 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2zwt\" (UniqueName: \"kubernetes.io/projected/13ae0f7e-b016-4044-b6db-4210a79854ab-kube-api-access-b2zwt\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.856570 4923 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1100409d-07f5-424b-8b9d-f8e8bd066a5c-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.856581 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13ae0f7e-b016-4044-b6db-4210a79854ab-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.856589 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1100409d-07f5-424b-8b9d-f8e8bd066a5c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.856599 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/13ae0f7e-b016-4044-b6db-4210a79854ab-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.856608 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpw2r\" (UniqueName: \"kubernetes.io/projected/1100409d-07f5-424b-8b9d-f8e8bd066a5c-kube-api-access-zpw2r\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:32 crc kubenswrapper[4923]: I0311 13:25:32.928160 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cqdq8" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.093599 4923 generic.go:334] "Generic (PLEG): container finished" podID="13ae0f7e-b016-4044-b6db-4210a79854ab" containerID="b7ad9e1425218d639b26e21c5bcb52eab894a864e3e4781633b36fb82ba8826f" exitCode=0 Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.093658 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" event={"ID":"13ae0f7e-b016-4044-b6db-4210a79854ab","Type":"ContainerDied","Data":"b7ad9e1425218d639b26e21c5bcb52eab894a864e3e4781633b36fb82ba8826f"} Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.093660 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.093690 4923 scope.go:117] "RemoveContainer" containerID="b7ad9e1425218d639b26e21c5bcb52eab894a864e3e4781633b36fb82ba8826f" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.093681 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk" event={"ID":"13ae0f7e-b016-4044-b6db-4210a79854ab","Type":"ContainerDied","Data":"57d2ae1d56b993a2af0f1e6a19ae7758401ceef3b2150b6a2f6b57c1c5a8e349"} Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.095118 4923 generic.go:334] "Generic (PLEG): container finished" podID="1100409d-07f5-424b-8b9d-f8e8bd066a5c" containerID="04ee7610f6b150a515d1dcae9289d9fcb3dcc44708dffead179693c530f02725" exitCode=0 Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.095700 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.096108 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" event={"ID":"1100409d-07f5-424b-8b9d-f8e8bd066a5c","Type":"ContainerDied","Data":"04ee7610f6b150a515d1dcae9289d9fcb3dcc44708dffead179693c530f02725"} Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.096143 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7f5b598bff-5r5gv" event={"ID":"1100409d-07f5-424b-8b9d-f8e8bd066a5c","Type":"ContainerDied","Data":"d2985020594cc12b88451cd0b8b01be331706c399d0533c5f0ee1b0bb6bb4ba2"} Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.113860 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk"] Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.117577 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5b758c8ff-btrzk"] Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.127401 4923 scope.go:117] "RemoveContainer" containerID="b7ad9e1425218d639b26e21c5bcb52eab894a864e3e4781633b36fb82ba8826f" Mar 11 13:25:33 crc kubenswrapper[4923]: E0311 13:25:33.128177 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7ad9e1425218d639b26e21c5bcb52eab894a864e3e4781633b36fb82ba8826f\": container with ID starting with b7ad9e1425218d639b26e21c5bcb52eab894a864e3e4781633b36fb82ba8826f not found: ID does not exist" containerID="b7ad9e1425218d639b26e21c5bcb52eab894a864e3e4781633b36fb82ba8826f" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.128240 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7ad9e1425218d639b26e21c5bcb52eab894a864e3e4781633b36fb82ba8826f"} err="failed to get container status \"b7ad9e1425218d639b26e21c5bcb52eab894a864e3e4781633b36fb82ba8826f\": rpc error: code = NotFound desc = could not find container \"b7ad9e1425218d639b26e21c5bcb52eab894a864e3e4781633b36fb82ba8826f\": container with ID starting with b7ad9e1425218d639b26e21c5bcb52eab894a864e3e4781633b36fb82ba8826f not found: ID does not exist" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.128269 4923 scope.go:117] "RemoveContainer" containerID="04ee7610f6b150a515d1dcae9289d9fcb3dcc44708dffead179693c530f02725" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.129694 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7f5b598bff-5r5gv"] Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.134662 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7f5b598bff-5r5gv"] Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.169254 4923 scope.go:117] "RemoveContainer" containerID="04ee7610f6b150a515d1dcae9289d9fcb3dcc44708dffead179693c530f02725" Mar 11 13:25:33 crc kubenswrapper[4923]: E0311 13:25:33.169737 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04ee7610f6b150a515d1dcae9289d9fcb3dcc44708dffead179693c530f02725\": container with ID starting with 04ee7610f6b150a515d1dcae9289d9fcb3dcc44708dffead179693c530f02725 not found: ID does not exist" containerID="04ee7610f6b150a515d1dcae9289d9fcb3dcc44708dffead179693c530f02725" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.169834 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04ee7610f6b150a515d1dcae9289d9fcb3dcc44708dffead179693c530f02725"} err="failed to get container status \"04ee7610f6b150a515d1dcae9289d9fcb3dcc44708dffead179693c530f02725\": rpc error: code = NotFound desc = could not find container \"04ee7610f6b150a515d1dcae9289d9fcb3dcc44708dffead179693c530f02725\": container with ID starting with 04ee7610f6b150a515d1dcae9289d9fcb3dcc44708dffead179693c530f02725 not found: ID does not exist" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.522567 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8"] Mar 11 13:25:33 crc kubenswrapper[4923]: E0311 13:25:33.522931 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bdd5510-3bb1-42f0-8646-c45f11bdeafe" containerName="extract-utilities" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.523317 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bdd5510-3bb1-42f0-8646-c45f11bdeafe" containerName="extract-utilities" Mar 11 13:25:33 crc kubenswrapper[4923]: E0311 13:25:33.523399 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1100409d-07f5-424b-8b9d-f8e8bd066a5c" containerName="controller-manager" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.523419 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="1100409d-07f5-424b-8b9d-f8e8bd066a5c" containerName="controller-manager" Mar 11 13:25:33 crc kubenswrapper[4923]: E0311 13:25:33.523446 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbba7a9b-2c9b-4770-ac96-b89390457eab" containerName="extract-content" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.523462 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbba7a9b-2c9b-4770-ac96-b89390457eab" containerName="extract-content" Mar 11 13:25:33 crc kubenswrapper[4923]: E0311 13:25:33.523478 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bdd5510-3bb1-42f0-8646-c45f11bdeafe" containerName="registry-server" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.523491 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bdd5510-3bb1-42f0-8646-c45f11bdeafe" containerName="registry-server" Mar 11 13:25:33 crc kubenswrapper[4923]: E0311 13:25:33.523508 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ab29276-3a67-42c9-8308-2ee71a3dd4dc" containerName="extract-utilities" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.523521 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ab29276-3a67-42c9-8308-2ee71a3dd4dc" containerName="extract-utilities" Mar 11 13:25:33 crc kubenswrapper[4923]: E0311 13:25:33.523544 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ab29276-3a67-42c9-8308-2ee71a3dd4dc" containerName="extract-content" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.523559 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ab29276-3a67-42c9-8308-2ee71a3dd4dc" containerName="extract-content" Mar 11 13:25:33 crc kubenswrapper[4923]: E0311 13:25:33.523577 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbba7a9b-2c9b-4770-ac96-b89390457eab" containerName="extract-utilities" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.523615 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbba7a9b-2c9b-4770-ac96-b89390457eab" containerName="extract-utilities" Mar 11 13:25:33 crc kubenswrapper[4923]: E0311 13:25:33.523641 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ab29276-3a67-42c9-8308-2ee71a3dd4dc" containerName="registry-server" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.523654 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ab29276-3a67-42c9-8308-2ee71a3dd4dc" containerName="registry-server" Mar 11 13:25:33 crc kubenswrapper[4923]: E0311 13:25:33.523669 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bdd5510-3bb1-42f0-8646-c45f11bdeafe" containerName="extract-content" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.523681 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bdd5510-3bb1-42f0-8646-c45f11bdeafe" containerName="extract-content" Mar 11 13:25:33 crc kubenswrapper[4923]: E0311 13:25:33.523699 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbba7a9b-2c9b-4770-ac96-b89390457eab" containerName="registry-server" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.523710 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbba7a9b-2c9b-4770-ac96-b89390457eab" containerName="registry-server" Mar 11 13:25:33 crc kubenswrapper[4923]: E0311 13:25:33.523725 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13ae0f7e-b016-4044-b6db-4210a79854ab" containerName="route-controller-manager" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.523740 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="13ae0f7e-b016-4044-b6db-4210a79854ab" containerName="route-controller-manager" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.523931 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="13ae0f7e-b016-4044-b6db-4210a79854ab" containerName="route-controller-manager" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.523949 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="1100409d-07f5-424b-8b9d-f8e8bd066a5c" containerName="controller-manager" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.523975 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ab29276-3a67-42c9-8308-2ee71a3dd4dc" containerName="registry-server" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.523994 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbba7a9b-2c9b-4770-ac96-b89390457eab" containerName="registry-server" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.524016 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bdd5510-3bb1-42f0-8646-c45f11bdeafe" containerName="registry-server" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.524669 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.528426 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.528427 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.528627 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.528655 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.528867 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.528992 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.534424 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-cd869d699-24j4g"] Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.535593 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.537265 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.540842 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.541143 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8"] Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.541272 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.541606 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.543818 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.544258 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.547523 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-cd869d699-24j4g"] Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.599915 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97c5q\" (UniqueName: \"kubernetes.io/projected/4082e161-7d1d-4041-a8c8-872ba90927eb-kube-api-access-97c5q\") pod \"route-controller-manager-6658d6fcd6-wvrv8\" (UID: \"4082e161-7d1d-4041-a8c8-872ba90927eb\") " pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.600059 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f6a7a4d4-f22e-4979-8289-9b754501d485-client-ca\") pod \"controller-manager-cd869d699-24j4g\" (UID: \"f6a7a4d4-f22e-4979-8289-9b754501d485\") " pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.600155 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4082e161-7d1d-4041-a8c8-872ba90927eb-serving-cert\") pod \"route-controller-manager-6658d6fcd6-wvrv8\" (UID: \"4082e161-7d1d-4041-a8c8-872ba90927eb\") " pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.600215 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4082e161-7d1d-4041-a8c8-872ba90927eb-config\") pod \"route-controller-manager-6658d6fcd6-wvrv8\" (UID: \"4082e161-7d1d-4041-a8c8-872ba90927eb\") " pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.600259 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4082e161-7d1d-4041-a8c8-872ba90927eb-client-ca\") pod \"route-controller-manager-6658d6fcd6-wvrv8\" (UID: \"4082e161-7d1d-4041-a8c8-872ba90927eb\") " pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.600666 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f6a7a4d4-f22e-4979-8289-9b754501d485-proxy-ca-bundles\") pod \"controller-manager-cd869d699-24j4g\" (UID: \"f6a7a4d4-f22e-4979-8289-9b754501d485\") " pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.600798 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv7v9\" (UniqueName: \"kubernetes.io/projected/f6a7a4d4-f22e-4979-8289-9b754501d485-kube-api-access-wv7v9\") pod \"controller-manager-cd869d699-24j4g\" (UID: \"f6a7a4d4-f22e-4979-8289-9b754501d485\") " pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.600913 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6a7a4d4-f22e-4979-8289-9b754501d485-serving-cert\") pod \"controller-manager-cd869d699-24j4g\" (UID: \"f6a7a4d4-f22e-4979-8289-9b754501d485\") " pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.600974 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6a7a4d4-f22e-4979-8289-9b754501d485-config\") pod \"controller-manager-cd869d699-24j4g\" (UID: \"f6a7a4d4-f22e-4979-8289-9b754501d485\") " pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.601990 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.702889 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv7v9\" (UniqueName: \"kubernetes.io/projected/f6a7a4d4-f22e-4979-8289-9b754501d485-kube-api-access-wv7v9\") pod \"controller-manager-cd869d699-24j4g\" (UID: \"f6a7a4d4-f22e-4979-8289-9b754501d485\") " pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.702975 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6a7a4d4-f22e-4979-8289-9b754501d485-serving-cert\") pod \"controller-manager-cd869d699-24j4g\" (UID: \"f6a7a4d4-f22e-4979-8289-9b754501d485\") " pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.703011 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6a7a4d4-f22e-4979-8289-9b754501d485-config\") pod \"controller-manager-cd869d699-24j4g\" (UID: \"f6a7a4d4-f22e-4979-8289-9b754501d485\") " pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.703059 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97c5q\" (UniqueName: \"kubernetes.io/projected/4082e161-7d1d-4041-a8c8-872ba90927eb-kube-api-access-97c5q\") pod \"route-controller-manager-6658d6fcd6-wvrv8\" (UID: \"4082e161-7d1d-4041-a8c8-872ba90927eb\") " pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.703089 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f6a7a4d4-f22e-4979-8289-9b754501d485-client-ca\") pod \"controller-manager-cd869d699-24j4g\" (UID: \"f6a7a4d4-f22e-4979-8289-9b754501d485\") " pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.703117 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4082e161-7d1d-4041-a8c8-872ba90927eb-serving-cert\") pod \"route-controller-manager-6658d6fcd6-wvrv8\" (UID: \"4082e161-7d1d-4041-a8c8-872ba90927eb\") " pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.703144 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4082e161-7d1d-4041-a8c8-872ba90927eb-config\") pod \"route-controller-manager-6658d6fcd6-wvrv8\" (UID: \"4082e161-7d1d-4041-a8c8-872ba90927eb\") " pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.703167 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4082e161-7d1d-4041-a8c8-872ba90927eb-client-ca\") pod \"route-controller-manager-6658d6fcd6-wvrv8\" (UID: \"4082e161-7d1d-4041-a8c8-872ba90927eb\") " pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.703206 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f6a7a4d4-f22e-4979-8289-9b754501d485-proxy-ca-bundles\") pod \"controller-manager-cd869d699-24j4g\" (UID: \"f6a7a4d4-f22e-4979-8289-9b754501d485\") " pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.704734 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f6a7a4d4-f22e-4979-8289-9b754501d485-proxy-ca-bundles\") pod \"controller-manager-cd869d699-24j4g\" (UID: \"f6a7a4d4-f22e-4979-8289-9b754501d485\") " pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.705302 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f6a7a4d4-f22e-4979-8289-9b754501d485-client-ca\") pod \"controller-manager-cd869d699-24j4g\" (UID: \"f6a7a4d4-f22e-4979-8289-9b754501d485\") " pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.706364 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6a7a4d4-f22e-4979-8289-9b754501d485-config\") pod \"controller-manager-cd869d699-24j4g\" (UID: \"f6a7a4d4-f22e-4979-8289-9b754501d485\") " pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.706377 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4082e161-7d1d-4041-a8c8-872ba90927eb-client-ca\") pod \"route-controller-manager-6658d6fcd6-wvrv8\" (UID: \"4082e161-7d1d-4041-a8c8-872ba90927eb\") " pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.707472 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4082e161-7d1d-4041-a8c8-872ba90927eb-config\") pod \"route-controller-manager-6658d6fcd6-wvrv8\" (UID: \"4082e161-7d1d-4041-a8c8-872ba90927eb\") " pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.718007 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4082e161-7d1d-4041-a8c8-872ba90927eb-serving-cert\") pod \"route-controller-manager-6658d6fcd6-wvrv8\" (UID: \"4082e161-7d1d-4041-a8c8-872ba90927eb\") " pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.718069 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6a7a4d4-f22e-4979-8289-9b754501d485-serving-cert\") pod \"controller-manager-cd869d699-24j4g\" (UID: \"f6a7a4d4-f22e-4979-8289-9b754501d485\") " pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.720690 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv7v9\" (UniqueName: \"kubernetes.io/projected/f6a7a4d4-f22e-4979-8289-9b754501d485-kube-api-access-wv7v9\") pod \"controller-manager-cd869d699-24j4g\" (UID: \"f6a7a4d4-f22e-4979-8289-9b754501d485\") " pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.726633 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97c5q\" (UniqueName: \"kubernetes.io/projected/4082e161-7d1d-4041-a8c8-872ba90927eb-kube-api-access-97c5q\") pod \"route-controller-manager-6658d6fcd6-wvrv8\" (UID: \"4082e161-7d1d-4041-a8c8-872ba90927eb\") " pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.928239 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" Mar 11 13:25:33 crc kubenswrapper[4923]: I0311 13:25:33.935279 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" Mar 11 13:25:34 crc kubenswrapper[4923]: I0311 13:25:34.419929 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8"] Mar 11 13:25:34 crc kubenswrapper[4923]: W0311 13:25:34.430000 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4082e161_7d1d_4041_a8c8_872ba90927eb.slice/crio-dc2a096e985f2473c90dd2d5f71d460f8664be417a0df9d8418b814a8cc127b8 WatchSource:0}: Error finding container dc2a096e985f2473c90dd2d5f71d460f8664be417a0df9d8418b814a8cc127b8: Status 404 returned error can't find the container with id dc2a096e985f2473c90dd2d5f71d460f8664be417a0df9d8418b814a8cc127b8 Mar 11 13:25:34 crc kubenswrapper[4923]: I0311 13:25:34.472362 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-cd869d699-24j4g"] Mar 11 13:25:34 crc kubenswrapper[4923]: W0311 13:25:34.480003 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf6a7a4d4_f22e_4979_8289_9b754501d485.slice/crio-644fc7f5a0dad954e7333a19497960e2f7851f7cc713a60a54a8995f6507ef73 WatchSource:0}: Error finding container 644fc7f5a0dad954e7333a19497960e2f7851f7cc713a60a54a8995f6507ef73: Status 404 returned error can't find the container with id 644fc7f5a0dad954e7333a19497960e2f7851f7cc713a60a54a8995f6507ef73 Mar 11 13:25:34 crc kubenswrapper[4923]: I0311 13:25:34.957545 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jcncp" Mar 11 13:25:35 crc kubenswrapper[4923]: I0311 13:25:35.006243 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jcncp" Mar 11 13:25:35 crc kubenswrapper[4923]: I0311 13:25:35.028500 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1100409d-07f5-424b-8b9d-f8e8bd066a5c" path="/var/lib/kubelet/pods/1100409d-07f5-424b-8b9d-f8e8bd066a5c/volumes" Mar 11 13:25:35 crc kubenswrapper[4923]: I0311 13:25:35.029368 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13ae0f7e-b016-4044-b6db-4210a79854ab" path="/var/lib/kubelet/pods/13ae0f7e-b016-4044-b6db-4210a79854ab/volumes" Mar 11 13:25:35 crc kubenswrapper[4923]: I0311 13:25:35.111776 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" event={"ID":"f6a7a4d4-f22e-4979-8289-9b754501d485","Type":"ContainerStarted","Data":"fbf2c77fb29afd47e7b74fe4e7733f4c43c3c6ccef5f2754de3b378f3c5b17c8"} Mar 11 13:25:35 crc kubenswrapper[4923]: I0311 13:25:35.111854 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" Mar 11 13:25:35 crc kubenswrapper[4923]: I0311 13:25:35.111869 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" event={"ID":"f6a7a4d4-f22e-4979-8289-9b754501d485","Type":"ContainerStarted","Data":"644fc7f5a0dad954e7333a19497960e2f7851f7cc713a60a54a8995f6507ef73"} Mar 11 13:25:35 crc kubenswrapper[4923]: I0311 13:25:35.112972 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" event={"ID":"4082e161-7d1d-4041-a8c8-872ba90927eb","Type":"ContainerStarted","Data":"0caabdb0b4f09a01fd095097d5d52a5d07f829e2521d327f2ea7ddf9e141f457"} Mar 11 13:25:35 crc kubenswrapper[4923]: I0311 13:25:35.113004 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" event={"ID":"4082e161-7d1d-4041-a8c8-872ba90927eb","Type":"ContainerStarted","Data":"dc2a096e985f2473c90dd2d5f71d460f8664be417a0df9d8418b814a8cc127b8"} Mar 11 13:25:35 crc kubenswrapper[4923]: I0311 13:25:35.117656 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" Mar 11 13:25:35 crc kubenswrapper[4923]: I0311 13:25:35.133160 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" podStartSLOduration=3.133143207 podStartE2EDuration="3.133143207s" podCreationTimestamp="2026-03-11 13:25:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:25:35.129482287 +0000 UTC m=+182.811384001" watchObservedRunningTime="2026-03-11 13:25:35.133143207 +0000 UTC m=+182.815044921" Mar 11 13:25:35 crc kubenswrapper[4923]: I0311 13:25:35.156994 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" podStartSLOduration=3.156976281 podStartE2EDuration="3.156976281s" podCreationTimestamp="2026-03-11 13:25:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:25:35.15327997 +0000 UTC m=+182.835181694" watchObservedRunningTime="2026-03-11 13:25:35.156976281 +0000 UTC m=+182.838877995" Mar 11 13:25:36 crc kubenswrapper[4923]: I0311 13:25:36.121399 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" Mar 11 13:25:36 crc kubenswrapper[4923]: I0311 13:25:36.132041 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.268773 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" podUID="08068104-34d8-440d-8f98-f9a6216af858" containerName="oauth-openshift" containerID="cri-o://d972361bb06146bbdc4aa442393491cba7ef1d4601e5da3040177b4bf8290720" gracePeriod=15 Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.785487 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.812756 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-serving-cert\") pod \"08068104-34d8-440d-8f98-f9a6216af858\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.814192 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-ocp-branding-template\") pod \"08068104-34d8-440d-8f98-f9a6216af858\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.814281 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-service-ca\") pod \"08068104-34d8-440d-8f98-f9a6216af858\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.814312 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/08068104-34d8-440d-8f98-f9a6216af858-audit-dir\") pod \"08068104-34d8-440d-8f98-f9a6216af858\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.814446 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-template-provider-selection\") pod \"08068104-34d8-440d-8f98-f9a6216af858\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.814499 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-router-certs\") pod \"08068104-34d8-440d-8f98-f9a6216af858\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.814496 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/08068104-34d8-440d-8f98-f9a6216af858-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "08068104-34d8-440d-8f98-f9a6216af858" (UID: "08068104-34d8-440d-8f98-f9a6216af858"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.814973 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "08068104-34d8-440d-8f98-f9a6216af858" (UID: "08068104-34d8-440d-8f98-f9a6216af858"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.814883 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-template-login\") pod \"08068104-34d8-440d-8f98-f9a6216af858\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.815231 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-session\") pod \"08068104-34d8-440d-8f98-f9a6216af858\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.815282 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-audit-policies\") pod \"08068104-34d8-440d-8f98-f9a6216af858\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.815311 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-idp-0-file-data\") pod \"08068104-34d8-440d-8f98-f9a6216af858\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.815389 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nx5b4\" (UniqueName: \"kubernetes.io/projected/08068104-34d8-440d-8f98-f9a6216af858-kube-api-access-nx5b4\") pod \"08068104-34d8-440d-8f98-f9a6216af858\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.815444 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-template-error\") pod \"08068104-34d8-440d-8f98-f9a6216af858\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.815477 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-trusted-ca-bundle\") pod \"08068104-34d8-440d-8f98-f9a6216af858\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.815527 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-cliconfig\") pod \"08068104-34d8-440d-8f98-f9a6216af858\" (UID: \"08068104-34d8-440d-8f98-f9a6216af858\") " Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.815941 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.815969 4923 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/08068104-34d8-440d-8f98-f9a6216af858-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.816722 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "08068104-34d8-440d-8f98-f9a6216af858" (UID: "08068104-34d8-440d-8f98-f9a6216af858"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.821811 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "08068104-34d8-440d-8f98-f9a6216af858" (UID: "08068104-34d8-440d-8f98-f9a6216af858"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.823789 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "08068104-34d8-440d-8f98-f9a6216af858" (UID: "08068104-34d8-440d-8f98-f9a6216af858"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.849519 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "08068104-34d8-440d-8f98-f9a6216af858" (UID: "08068104-34d8-440d-8f98-f9a6216af858"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.850154 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "08068104-34d8-440d-8f98-f9a6216af858" (UID: "08068104-34d8-440d-8f98-f9a6216af858"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.851958 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "08068104-34d8-440d-8f98-f9a6216af858" (UID: "08068104-34d8-440d-8f98-f9a6216af858"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.854660 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08068104-34d8-440d-8f98-f9a6216af858-kube-api-access-nx5b4" (OuterVolumeSpecName: "kube-api-access-nx5b4") pod "08068104-34d8-440d-8f98-f9a6216af858" (UID: "08068104-34d8-440d-8f98-f9a6216af858"). InnerVolumeSpecName "kube-api-access-nx5b4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.854741 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "08068104-34d8-440d-8f98-f9a6216af858" (UID: "08068104-34d8-440d-8f98-f9a6216af858"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.854956 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "08068104-34d8-440d-8f98-f9a6216af858" (UID: "08068104-34d8-440d-8f98-f9a6216af858"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.855316 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "08068104-34d8-440d-8f98-f9a6216af858" (UID: "08068104-34d8-440d-8f98-f9a6216af858"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.855564 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "08068104-34d8-440d-8f98-f9a6216af858" (UID: "08068104-34d8-440d-8f98-f9a6216af858"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.856771 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "08068104-34d8-440d-8f98-f9a6216af858" (UID: "08068104-34d8-440d-8f98-f9a6216af858"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.917317 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.917386 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.917404 4923 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.917418 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.917431 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nx5b4\" (UniqueName: \"kubernetes.io/projected/08068104-34d8-440d-8f98-f9a6216af858-kube-api-access-nx5b4\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.917443 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.917454 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.917466 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.917478 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.917489 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.917501 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:40 crc kubenswrapper[4923]: I0311 13:25:40.917515 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/08068104-34d8-440d-8f98-f9a6216af858-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:41 crc kubenswrapper[4923]: I0311 13:25:41.156334 4923 generic.go:334] "Generic (PLEG): container finished" podID="08068104-34d8-440d-8f98-f9a6216af858" containerID="d972361bb06146bbdc4aa442393491cba7ef1d4601e5da3040177b4bf8290720" exitCode=0 Mar 11 13:25:41 crc kubenswrapper[4923]: I0311 13:25:41.156447 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" event={"ID":"08068104-34d8-440d-8f98-f9a6216af858","Type":"ContainerDied","Data":"d972361bb06146bbdc4aa442393491cba7ef1d4601e5da3040177b4bf8290720"} Mar 11 13:25:41 crc kubenswrapper[4923]: I0311 13:25:41.156488 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" Mar 11 13:25:41 crc kubenswrapper[4923]: I0311 13:25:41.156555 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-hgkc2" event={"ID":"08068104-34d8-440d-8f98-f9a6216af858","Type":"ContainerDied","Data":"41074d26a93e59829551b8f1b148bd83529020e9b7f09d8e91e43fcb0c06c711"} Mar 11 13:25:41 crc kubenswrapper[4923]: I0311 13:25:41.156598 4923 scope.go:117] "RemoveContainer" containerID="d972361bb06146bbdc4aa442393491cba7ef1d4601e5da3040177b4bf8290720" Mar 11 13:25:41 crc kubenswrapper[4923]: I0311 13:25:41.174567 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hgkc2"] Mar 11 13:25:41 crc kubenswrapper[4923]: I0311 13:25:41.176947 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-hgkc2"] Mar 11 13:25:41 crc kubenswrapper[4923]: I0311 13:25:41.182196 4923 scope.go:117] "RemoveContainer" containerID="d972361bb06146bbdc4aa442393491cba7ef1d4601e5da3040177b4bf8290720" Mar 11 13:25:41 crc kubenswrapper[4923]: E0311 13:25:41.182908 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d972361bb06146bbdc4aa442393491cba7ef1d4601e5da3040177b4bf8290720\": container with ID starting with d972361bb06146bbdc4aa442393491cba7ef1d4601e5da3040177b4bf8290720 not found: ID does not exist" containerID="d972361bb06146bbdc4aa442393491cba7ef1d4601e5da3040177b4bf8290720" Mar 11 13:25:41 crc kubenswrapper[4923]: I0311 13:25:41.182946 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d972361bb06146bbdc4aa442393491cba7ef1d4601e5da3040177b4bf8290720"} err="failed to get container status \"d972361bb06146bbdc4aa442393491cba7ef1d4601e5da3040177b4bf8290720\": rpc error: code = NotFound desc = could not find container \"d972361bb06146bbdc4aa442393491cba7ef1d4601e5da3040177b4bf8290720\": container with ID starting with d972361bb06146bbdc4aa442393491cba7ef1d4601e5da3040177b4bf8290720 not found: ID does not exist" Mar 11 13:25:43 crc kubenswrapper[4923]: I0311 13:25:43.030275 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08068104-34d8-440d-8f98-f9a6216af858" path="/var/lib/kubelet/pods/08068104-34d8-440d-8f98-f9a6216af858/volumes" Mar 11 13:25:43 crc kubenswrapper[4923]: I0311 13:25:43.080418 4923 ???:1] "http: TLS handshake error from 192.168.126.11:42662: no serving certificate available for the kubelet" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.530128 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-57bcd9fbb-w86mb"] Mar 11 13:25:49 crc kubenswrapper[4923]: E0311 13:25:49.530925 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08068104-34d8-440d-8f98-f9a6216af858" containerName="oauth-openshift" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.530943 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="08068104-34d8-440d-8f98-f9a6216af858" containerName="oauth-openshift" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.531068 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="08068104-34d8-440d-8f98-f9a6216af858" containerName="oauth-openshift" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.531509 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.534734 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.537746 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.538033 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.540964 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-57bcd9fbb-w86mb"] Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.541925 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.542281 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.542753 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.543011 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.543068 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.543026 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.543389 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.543617 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.547222 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.556666 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.563450 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.572964 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.634631 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8c039ace-f501-4acc-8b3f-d0e22a485742-audit-dir\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.634691 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.634725 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.634763 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-system-router-certs\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.634805 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-system-session\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.634879 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-system-service-ca\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.634908 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.634932 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8c039ace-f501-4acc-8b3f-d0e22a485742-audit-policies\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.634955 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-user-template-error\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.634997 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-user-template-login\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.635191 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.635289 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.635382 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkxpc\" (UniqueName: \"kubernetes.io/projected/8c039ace-f501-4acc-8b3f-d0e22a485742-kube-api-access-pkxpc\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.635432 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.737148 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.737226 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-user-template-error\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.737265 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8c039ace-f501-4acc-8b3f-d0e22a485742-audit-policies\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.737304 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-user-template-login\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.737404 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.737456 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.737501 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkxpc\" (UniqueName: \"kubernetes.io/projected/8c039ace-f501-4acc-8b3f-d0e22a485742-kube-api-access-pkxpc\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.737545 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.737588 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8c039ace-f501-4acc-8b3f-d0e22a485742-audit-dir\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.737624 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.737660 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.737708 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-system-router-certs\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.737771 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-system-session\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.737813 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-system-service-ca\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.739146 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/8c039ace-f501-4acc-8b3f-d0e22a485742-audit-dir\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.740079 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-system-cliconfig\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.740175 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-system-service-ca\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.740608 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/8c039ace-f501-4acc-8b3f-d0e22a485742-audit-policies\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.743404 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.744880 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-system-router-certs\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.745842 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.746268 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-system-serving-cert\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.747111 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-user-template-error\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.747566 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-user-template-login\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.749107 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-system-session\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.752209 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.757405 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/8c039ace-f501-4acc-8b3f-d0e22a485742-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.763964 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkxpc\" (UniqueName: \"kubernetes.io/projected/8c039ace-f501-4acc-8b3f-d0e22a485742-kube-api-access-pkxpc\") pod \"oauth-openshift-57bcd9fbb-w86mb\" (UID: \"8c039ace-f501-4acc-8b3f-d0e22a485742\") " pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:49 crc kubenswrapper[4923]: I0311 13:25:49.862794 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:50 crc kubenswrapper[4923]: I0311 13:25:50.418020 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-57bcd9fbb-w86mb"] Mar 11 13:25:51 crc kubenswrapper[4923]: I0311 13:25:51.220552 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" event={"ID":"8c039ace-f501-4acc-8b3f-d0e22a485742","Type":"ContainerStarted","Data":"9b357fcbde23b02699963fb1d1437fc95e7fdcf607c8f327b699c8ce318ecb33"} Mar 11 13:25:51 crc kubenswrapper[4923]: I0311 13:25:51.220930 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:51 crc kubenswrapper[4923]: I0311 13:25:51.220968 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" event={"ID":"8c039ace-f501-4acc-8b3f-d0e22a485742","Type":"ContainerStarted","Data":"7de2f9ec3ea08b198a51334af32e8b70f82ab2c18a5905f58755423f43e1eed0"} Mar 11 13:25:51 crc kubenswrapper[4923]: I0311 13:25:51.235896 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" Mar 11 13:25:51 crc kubenswrapper[4923]: I0311 13:25:51.303181 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-57bcd9fbb-w86mb" podStartSLOduration=36.30314497 podStartE2EDuration="36.30314497s" podCreationTimestamp="2026-03-11 13:25:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:25:51.256799918 +0000 UTC m=+198.938701672" watchObservedRunningTime="2026-03-11 13:25:51.30314497 +0000 UTC m=+198.985046734" Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.103228 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-cd869d699-24j4g"] Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.103629 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" podUID="f6a7a4d4-f22e-4979-8289-9b754501d485" containerName="controller-manager" containerID="cri-o://fbf2c77fb29afd47e7b74fe4e7733f4c43c3c6ccef5f2754de3b378f3c5b17c8" gracePeriod=30 Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.205546 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8"] Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.205857 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" podUID="4082e161-7d1d-4041-a8c8-872ba90927eb" containerName="route-controller-manager" containerID="cri-o://0caabdb0b4f09a01fd095097d5d52a5d07f829e2521d327f2ea7ddf9e141f457" gracePeriod=30 Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.230084 4923 generic.go:334] "Generic (PLEG): container finished" podID="f6a7a4d4-f22e-4979-8289-9b754501d485" containerID="fbf2c77fb29afd47e7b74fe4e7733f4c43c3c6ccef5f2754de3b378f3c5b17c8" exitCode=0 Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.230504 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" event={"ID":"f6a7a4d4-f22e-4979-8289-9b754501d485","Type":"ContainerDied","Data":"fbf2c77fb29afd47e7b74fe4e7733f4c43c3c6ccef5f2754de3b378f3c5b17c8"} Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.736855 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.779154 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4082e161-7d1d-4041-a8c8-872ba90927eb-config\") pod \"4082e161-7d1d-4041-a8c8-872ba90927eb\" (UID: \"4082e161-7d1d-4041-a8c8-872ba90927eb\") " Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.779204 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4082e161-7d1d-4041-a8c8-872ba90927eb-client-ca\") pod \"4082e161-7d1d-4041-a8c8-872ba90927eb\" (UID: \"4082e161-7d1d-4041-a8c8-872ba90927eb\") " Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.779256 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97c5q\" (UniqueName: \"kubernetes.io/projected/4082e161-7d1d-4041-a8c8-872ba90927eb-kube-api-access-97c5q\") pod \"4082e161-7d1d-4041-a8c8-872ba90927eb\" (UID: \"4082e161-7d1d-4041-a8c8-872ba90927eb\") " Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.779273 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4082e161-7d1d-4041-a8c8-872ba90927eb-serving-cert\") pod \"4082e161-7d1d-4041-a8c8-872ba90927eb\" (UID: \"4082e161-7d1d-4041-a8c8-872ba90927eb\") " Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.780414 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4082e161-7d1d-4041-a8c8-872ba90927eb-client-ca" (OuterVolumeSpecName: "client-ca") pod "4082e161-7d1d-4041-a8c8-872ba90927eb" (UID: "4082e161-7d1d-4041-a8c8-872ba90927eb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.780449 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4082e161-7d1d-4041-a8c8-872ba90927eb-config" (OuterVolumeSpecName: "config") pod "4082e161-7d1d-4041-a8c8-872ba90927eb" (UID: "4082e161-7d1d-4041-a8c8-872ba90927eb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.785106 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4082e161-7d1d-4041-a8c8-872ba90927eb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4082e161-7d1d-4041-a8c8-872ba90927eb" (UID: "4082e161-7d1d-4041-a8c8-872ba90927eb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.787129 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4082e161-7d1d-4041-a8c8-872ba90927eb-kube-api-access-97c5q" (OuterVolumeSpecName: "kube-api-access-97c5q") pod "4082e161-7d1d-4041-a8c8-872ba90927eb" (UID: "4082e161-7d1d-4041-a8c8-872ba90927eb"). InnerVolumeSpecName "kube-api-access-97c5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.803063 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.880581 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f6a7a4d4-f22e-4979-8289-9b754501d485-proxy-ca-bundles\") pod \"f6a7a4d4-f22e-4979-8289-9b754501d485\" (UID: \"f6a7a4d4-f22e-4979-8289-9b754501d485\") " Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.880658 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wv7v9\" (UniqueName: \"kubernetes.io/projected/f6a7a4d4-f22e-4979-8289-9b754501d485-kube-api-access-wv7v9\") pod \"f6a7a4d4-f22e-4979-8289-9b754501d485\" (UID: \"f6a7a4d4-f22e-4979-8289-9b754501d485\") " Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.880677 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6a7a4d4-f22e-4979-8289-9b754501d485-serving-cert\") pod \"f6a7a4d4-f22e-4979-8289-9b754501d485\" (UID: \"f6a7a4d4-f22e-4979-8289-9b754501d485\") " Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.880701 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f6a7a4d4-f22e-4979-8289-9b754501d485-client-ca\") pod \"f6a7a4d4-f22e-4979-8289-9b754501d485\" (UID: \"f6a7a4d4-f22e-4979-8289-9b754501d485\") " Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.880736 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6a7a4d4-f22e-4979-8289-9b754501d485-config\") pod \"f6a7a4d4-f22e-4979-8289-9b754501d485\" (UID: \"f6a7a4d4-f22e-4979-8289-9b754501d485\") " Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.880929 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4082e161-7d1d-4041-a8c8-872ba90927eb-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.880939 4923 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4082e161-7d1d-4041-a8c8-872ba90927eb-client-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.880949 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97c5q\" (UniqueName: \"kubernetes.io/projected/4082e161-7d1d-4041-a8c8-872ba90927eb-kube-api-access-97c5q\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.880958 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4082e161-7d1d-4041-a8c8-872ba90927eb-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.881325 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6a7a4d4-f22e-4979-8289-9b754501d485-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "f6a7a4d4-f22e-4979-8289-9b754501d485" (UID: "f6a7a4d4-f22e-4979-8289-9b754501d485"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.881436 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6a7a4d4-f22e-4979-8289-9b754501d485-config" (OuterVolumeSpecName: "config") pod "f6a7a4d4-f22e-4979-8289-9b754501d485" (UID: "f6a7a4d4-f22e-4979-8289-9b754501d485"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.881458 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6a7a4d4-f22e-4979-8289-9b754501d485-client-ca" (OuterVolumeSpecName: "client-ca") pod "f6a7a4d4-f22e-4979-8289-9b754501d485" (UID: "f6a7a4d4-f22e-4979-8289-9b754501d485"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.884410 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6a7a4d4-f22e-4979-8289-9b754501d485-kube-api-access-wv7v9" (OuterVolumeSpecName: "kube-api-access-wv7v9") pod "f6a7a4d4-f22e-4979-8289-9b754501d485" (UID: "f6a7a4d4-f22e-4979-8289-9b754501d485"). InnerVolumeSpecName "kube-api-access-wv7v9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.885455 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6a7a4d4-f22e-4979-8289-9b754501d485-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f6a7a4d4-f22e-4979-8289-9b754501d485" (UID: "f6a7a4d4-f22e-4979-8289-9b754501d485"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.982231 4923 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f6a7a4d4-f22e-4979-8289-9b754501d485-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.982277 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wv7v9\" (UniqueName: \"kubernetes.io/projected/f6a7a4d4-f22e-4979-8289-9b754501d485-kube-api-access-wv7v9\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.982293 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f6a7a4d4-f22e-4979-8289-9b754501d485-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.982307 4923 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f6a7a4d4-f22e-4979-8289-9b754501d485-client-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:52 crc kubenswrapper[4923]: I0311 13:25:52.982318 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6a7a4d4-f22e-4979-8289-9b754501d485-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.237300 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" event={"ID":"f6a7a4d4-f22e-4979-8289-9b754501d485","Type":"ContainerDied","Data":"644fc7f5a0dad954e7333a19497960e2f7851f7cc713a60a54a8995f6507ef73"} Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.237437 4923 scope.go:117] "RemoveContainer" containerID="fbf2c77fb29afd47e7b74fe4e7733f4c43c3c6ccef5f2754de3b378f3c5b17c8" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.237867 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-cd869d699-24j4g" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.239029 4923 generic.go:334] "Generic (PLEG): container finished" podID="4082e161-7d1d-4041-a8c8-872ba90927eb" containerID="0caabdb0b4f09a01fd095097d5d52a5d07f829e2521d327f2ea7ddf9e141f457" exitCode=0 Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.239826 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.240242 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" event={"ID":"4082e161-7d1d-4041-a8c8-872ba90927eb","Type":"ContainerDied","Data":"0caabdb0b4f09a01fd095097d5d52a5d07f829e2521d327f2ea7ddf9e141f457"} Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.240270 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8" event={"ID":"4082e161-7d1d-4041-a8c8-872ba90927eb","Type":"ContainerDied","Data":"dc2a096e985f2473c90dd2d5f71d460f8664be417a0df9d8418b814a8cc127b8"} Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.259892 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8"] Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.261061 4923 scope.go:117] "RemoveContainer" containerID="0caabdb0b4f09a01fd095097d5d52a5d07f829e2521d327f2ea7ddf9e141f457" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.268092 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6658d6fcd6-wvrv8"] Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.278427 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-cd869d699-24j4g"] Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.280609 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-cd869d699-24j4g"] Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.281481 4923 scope.go:117] "RemoveContainer" containerID="0caabdb0b4f09a01fd095097d5d52a5d07f829e2521d327f2ea7ddf9e141f457" Mar 11 13:25:53 crc kubenswrapper[4923]: E0311 13:25:53.281922 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0caabdb0b4f09a01fd095097d5d52a5d07f829e2521d327f2ea7ddf9e141f457\": container with ID starting with 0caabdb0b4f09a01fd095097d5d52a5d07f829e2521d327f2ea7ddf9e141f457 not found: ID does not exist" containerID="0caabdb0b4f09a01fd095097d5d52a5d07f829e2521d327f2ea7ddf9e141f457" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.281961 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0caabdb0b4f09a01fd095097d5d52a5d07f829e2521d327f2ea7ddf9e141f457"} err="failed to get container status \"0caabdb0b4f09a01fd095097d5d52a5d07f829e2521d327f2ea7ddf9e141f457\": rpc error: code = NotFound desc = could not find container \"0caabdb0b4f09a01fd095097d5d52a5d07f829e2521d327f2ea7ddf9e141f457\": container with ID starting with 0caabdb0b4f09a01fd095097d5d52a5d07f829e2521d327f2ea7ddf9e141f457 not found: ID does not exist" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.536264 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75d8cbb545-xmbff"] Mar 11 13:25:53 crc kubenswrapper[4923]: E0311 13:25:53.536980 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4082e161-7d1d-4041-a8c8-872ba90927eb" containerName="route-controller-manager" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.537010 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="4082e161-7d1d-4041-a8c8-872ba90927eb" containerName="route-controller-manager" Mar 11 13:25:53 crc kubenswrapper[4923]: E0311 13:25:53.537058 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6a7a4d4-f22e-4979-8289-9b754501d485" containerName="controller-manager" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.537074 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6a7a4d4-f22e-4979-8289-9b754501d485" containerName="controller-manager" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.537266 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6a7a4d4-f22e-4979-8289-9b754501d485" containerName="controller-manager" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.537298 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="4082e161-7d1d-4041-a8c8-872ba90927eb" containerName="route-controller-manager" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.539288 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75d8cbb545-xmbff" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.539798 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6888d76f79-wzgcr"] Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.540576 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6888d76f79-wzgcr" Mar 11 13:25:53 crc kubenswrapper[4923]: W0311 13:25:53.544902 4923 reflector.go:561] object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2": failed to list *v1.Secret: secrets "route-controller-manager-sa-dockercfg-h2zr2" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Mar 11 13:25:53 crc kubenswrapper[4923]: E0311 13:25:53.544944 4923 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"route-controller-manager-sa-dockercfg-h2zr2\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"route-controller-manager-sa-dockercfg-h2zr2\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.545043 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.545058 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.545215 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 11 13:25:53 crc kubenswrapper[4923]: W0311 13:25:53.545383 4923 reflector.go:561] object-"openshift-route-controller-manager"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Mar 11 13:25:53 crc kubenswrapper[4923]: E0311 13:25:53.545408 4923 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.545534 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.547199 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.547974 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.548193 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.548582 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.548770 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.549507 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.554452 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75d8cbb545-xmbff"] Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.557098 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.566963 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6888d76f79-wzgcr"] Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.589963 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdshp\" (UniqueName: \"kubernetes.io/projected/d8f54ed0-dd9d-4cac-b00d-c8c378338a2e-kube-api-access-rdshp\") pod \"controller-manager-6888d76f79-wzgcr\" (UID: \"d8f54ed0-dd9d-4cac-b00d-c8c378338a2e\") " pod="openshift-controller-manager/controller-manager-6888d76f79-wzgcr" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.590038 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8f54ed0-dd9d-4cac-b00d-c8c378338a2e-client-ca\") pod \"controller-manager-6888d76f79-wzgcr\" (UID: \"d8f54ed0-dd9d-4cac-b00d-c8c378338a2e\") " pod="openshift-controller-manager/controller-manager-6888d76f79-wzgcr" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.590075 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57ac6c22-6124-437d-bbe7-59c376badfc4-serving-cert\") pod \"route-controller-manager-75d8cbb545-xmbff\" (UID: \"57ac6c22-6124-437d-bbe7-59c376badfc4\") " pod="openshift-route-controller-manager/route-controller-manager-75d8cbb545-xmbff" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.590098 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhs4x\" (UniqueName: \"kubernetes.io/projected/57ac6c22-6124-437d-bbe7-59c376badfc4-kube-api-access-zhs4x\") pod \"route-controller-manager-75d8cbb545-xmbff\" (UID: \"57ac6c22-6124-437d-bbe7-59c376badfc4\") " pod="openshift-route-controller-manager/route-controller-manager-75d8cbb545-xmbff" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.590141 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8f54ed0-dd9d-4cac-b00d-c8c378338a2e-serving-cert\") pod \"controller-manager-6888d76f79-wzgcr\" (UID: \"d8f54ed0-dd9d-4cac-b00d-c8c378338a2e\") " pod="openshift-controller-manager/controller-manager-6888d76f79-wzgcr" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.590166 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d8f54ed0-dd9d-4cac-b00d-c8c378338a2e-proxy-ca-bundles\") pod \"controller-manager-6888d76f79-wzgcr\" (UID: \"d8f54ed0-dd9d-4cac-b00d-c8c378338a2e\") " pod="openshift-controller-manager/controller-manager-6888d76f79-wzgcr" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.590192 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8f54ed0-dd9d-4cac-b00d-c8c378338a2e-config\") pod \"controller-manager-6888d76f79-wzgcr\" (UID: \"d8f54ed0-dd9d-4cac-b00d-c8c378338a2e\") " pod="openshift-controller-manager/controller-manager-6888d76f79-wzgcr" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.590214 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57ac6c22-6124-437d-bbe7-59c376badfc4-config\") pod \"route-controller-manager-75d8cbb545-xmbff\" (UID: \"57ac6c22-6124-437d-bbe7-59c376badfc4\") " pod="openshift-route-controller-manager/route-controller-manager-75d8cbb545-xmbff" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.590243 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/57ac6c22-6124-437d-bbe7-59c376badfc4-client-ca\") pod \"route-controller-manager-75d8cbb545-xmbff\" (UID: \"57ac6c22-6124-437d-bbe7-59c376badfc4\") " pod="openshift-route-controller-manager/route-controller-manager-75d8cbb545-xmbff" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.691763 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57ac6c22-6124-437d-bbe7-59c376badfc4-serving-cert\") pod \"route-controller-manager-75d8cbb545-xmbff\" (UID: \"57ac6c22-6124-437d-bbe7-59c376badfc4\") " pod="openshift-route-controller-manager/route-controller-manager-75d8cbb545-xmbff" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.691820 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhs4x\" (UniqueName: \"kubernetes.io/projected/57ac6c22-6124-437d-bbe7-59c376badfc4-kube-api-access-zhs4x\") pod \"route-controller-manager-75d8cbb545-xmbff\" (UID: \"57ac6c22-6124-437d-bbe7-59c376badfc4\") " pod="openshift-route-controller-manager/route-controller-manager-75d8cbb545-xmbff" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.691867 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8f54ed0-dd9d-4cac-b00d-c8c378338a2e-serving-cert\") pod \"controller-manager-6888d76f79-wzgcr\" (UID: \"d8f54ed0-dd9d-4cac-b00d-c8c378338a2e\") " pod="openshift-controller-manager/controller-manager-6888d76f79-wzgcr" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.691891 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d8f54ed0-dd9d-4cac-b00d-c8c378338a2e-proxy-ca-bundles\") pod \"controller-manager-6888d76f79-wzgcr\" (UID: \"d8f54ed0-dd9d-4cac-b00d-c8c378338a2e\") " pod="openshift-controller-manager/controller-manager-6888d76f79-wzgcr" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.691918 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8f54ed0-dd9d-4cac-b00d-c8c378338a2e-config\") pod \"controller-manager-6888d76f79-wzgcr\" (UID: \"d8f54ed0-dd9d-4cac-b00d-c8c378338a2e\") " pod="openshift-controller-manager/controller-manager-6888d76f79-wzgcr" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.691941 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57ac6c22-6124-437d-bbe7-59c376badfc4-config\") pod \"route-controller-manager-75d8cbb545-xmbff\" (UID: \"57ac6c22-6124-437d-bbe7-59c376badfc4\") " pod="openshift-route-controller-manager/route-controller-manager-75d8cbb545-xmbff" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.691968 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/57ac6c22-6124-437d-bbe7-59c376badfc4-client-ca\") pod \"route-controller-manager-75d8cbb545-xmbff\" (UID: \"57ac6c22-6124-437d-bbe7-59c376badfc4\") " pod="openshift-route-controller-manager/route-controller-manager-75d8cbb545-xmbff" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.692008 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdshp\" (UniqueName: \"kubernetes.io/projected/d8f54ed0-dd9d-4cac-b00d-c8c378338a2e-kube-api-access-rdshp\") pod \"controller-manager-6888d76f79-wzgcr\" (UID: \"d8f54ed0-dd9d-4cac-b00d-c8c378338a2e\") " pod="openshift-controller-manager/controller-manager-6888d76f79-wzgcr" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.692038 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8f54ed0-dd9d-4cac-b00d-c8c378338a2e-client-ca\") pod \"controller-manager-6888d76f79-wzgcr\" (UID: \"d8f54ed0-dd9d-4cac-b00d-c8c378338a2e\") " pod="openshift-controller-manager/controller-manager-6888d76f79-wzgcr" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.693074 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d8f54ed0-dd9d-4cac-b00d-c8c378338a2e-client-ca\") pod \"controller-manager-6888d76f79-wzgcr\" (UID: \"d8f54ed0-dd9d-4cac-b00d-c8c378338a2e\") " pod="openshift-controller-manager/controller-manager-6888d76f79-wzgcr" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.694375 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d8f54ed0-dd9d-4cac-b00d-c8c378338a2e-proxy-ca-bundles\") pod \"controller-manager-6888d76f79-wzgcr\" (UID: \"d8f54ed0-dd9d-4cac-b00d-c8c378338a2e\") " pod="openshift-controller-manager/controller-manager-6888d76f79-wzgcr" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.695203 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/57ac6c22-6124-437d-bbe7-59c376badfc4-client-ca\") pod \"route-controller-manager-75d8cbb545-xmbff\" (UID: \"57ac6c22-6124-437d-bbe7-59c376badfc4\") " pod="openshift-route-controller-manager/route-controller-manager-75d8cbb545-xmbff" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.696185 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/57ac6c22-6124-437d-bbe7-59c376badfc4-config\") pod \"route-controller-manager-75d8cbb545-xmbff\" (UID: \"57ac6c22-6124-437d-bbe7-59c376badfc4\") " pod="openshift-route-controller-manager/route-controller-manager-75d8cbb545-xmbff" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.696469 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d8f54ed0-dd9d-4cac-b00d-c8c378338a2e-config\") pod \"controller-manager-6888d76f79-wzgcr\" (UID: \"d8f54ed0-dd9d-4cac-b00d-c8c378338a2e\") " pod="openshift-controller-manager/controller-manager-6888d76f79-wzgcr" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.699073 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d8f54ed0-dd9d-4cac-b00d-c8c378338a2e-serving-cert\") pod \"controller-manager-6888d76f79-wzgcr\" (UID: \"d8f54ed0-dd9d-4cac-b00d-c8c378338a2e\") " pod="openshift-controller-manager/controller-manager-6888d76f79-wzgcr" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.699225 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/57ac6c22-6124-437d-bbe7-59c376badfc4-serving-cert\") pod \"route-controller-manager-75d8cbb545-xmbff\" (UID: \"57ac6c22-6124-437d-bbe7-59c376badfc4\") " pod="openshift-route-controller-manager/route-controller-manager-75d8cbb545-xmbff" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.713093 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdshp\" (UniqueName: \"kubernetes.io/projected/d8f54ed0-dd9d-4cac-b00d-c8c378338a2e-kube-api-access-rdshp\") pod \"controller-manager-6888d76f79-wzgcr\" (UID: \"d8f54ed0-dd9d-4cac-b00d-c8c378338a2e\") " pod="openshift-controller-manager/controller-manager-6888d76f79-wzgcr" Mar 11 13:25:53 crc kubenswrapper[4923]: I0311 13:25:53.871493 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6888d76f79-wzgcr" Mar 11 13:25:54 crc kubenswrapper[4923]: I0311 13:25:54.291762 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6888d76f79-wzgcr"] Mar 11 13:25:54 crc kubenswrapper[4923]: W0311 13:25:54.306515 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8f54ed0_dd9d_4cac_b00d_c8c378338a2e.slice/crio-3434dcd00a530133ffbbe1d21644b6c66b29131cae09c302d175fc52adf15929 WatchSource:0}: Error finding container 3434dcd00a530133ffbbe1d21644b6c66b29131cae09c302d175fc52adf15929: Status 404 returned error can't find the container with id 3434dcd00a530133ffbbe1d21644b6c66b29131cae09c302d175fc52adf15929 Mar 11 13:25:54 crc kubenswrapper[4923]: I0311 13:25:54.619732 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 11 13:25:54 crc kubenswrapper[4923]: I0311 13:25:54.702152 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 11 13:25:54 crc kubenswrapper[4923]: I0311 13:25:54.714860 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhs4x\" (UniqueName: \"kubernetes.io/projected/57ac6c22-6124-437d-bbe7-59c376badfc4-kube-api-access-zhs4x\") pod \"route-controller-manager-75d8cbb545-xmbff\" (UID: \"57ac6c22-6124-437d-bbe7-59c376badfc4\") " pod="openshift-route-controller-manager/route-controller-manager-75d8cbb545-xmbff" Mar 11 13:25:54 crc kubenswrapper[4923]: I0311 13:25:54.789108 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-75d8cbb545-xmbff" Mar 11 13:25:55 crc kubenswrapper[4923]: I0311 13:25:55.030028 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4082e161-7d1d-4041-a8c8-872ba90927eb" path="/var/lib/kubelet/pods/4082e161-7d1d-4041-a8c8-872ba90927eb/volumes" Mar 11 13:25:55 crc kubenswrapper[4923]: I0311 13:25:55.031047 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6a7a4d4-f22e-4979-8289-9b754501d485" path="/var/lib/kubelet/pods/f6a7a4d4-f22e-4979-8289-9b754501d485/volumes" Mar 11 13:25:55 crc kubenswrapper[4923]: I0311 13:25:55.192336 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-75d8cbb545-xmbff"] Mar 11 13:25:55 crc kubenswrapper[4923]: I0311 13:25:55.256764 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6888d76f79-wzgcr" event={"ID":"d8f54ed0-dd9d-4cac-b00d-c8c378338a2e","Type":"ContainerStarted","Data":"6a440307cec00b70c602f3c5a74767a700916e59841b6c65501372b7e2134a60"} Mar 11 13:25:55 crc kubenswrapper[4923]: I0311 13:25:55.257131 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6888d76f79-wzgcr" event={"ID":"d8f54ed0-dd9d-4cac-b00d-c8c378338a2e","Type":"ContainerStarted","Data":"3434dcd00a530133ffbbe1d21644b6c66b29131cae09c302d175fc52adf15929"} Mar 11 13:25:55 crc kubenswrapper[4923]: I0311 13:25:55.257531 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6888d76f79-wzgcr" Mar 11 13:25:55 crc kubenswrapper[4923]: I0311 13:25:55.258411 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75d8cbb545-xmbff" event={"ID":"57ac6c22-6124-437d-bbe7-59c376badfc4","Type":"ContainerStarted","Data":"0bb0fb52f683c85472431af589d4fadf9eafa59924734cfb4d2bbcea298107de"} Mar 11 13:25:55 crc kubenswrapper[4923]: I0311 13:25:55.263139 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6888d76f79-wzgcr" Mar 11 13:25:55 crc kubenswrapper[4923]: I0311 13:25:55.276459 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6888d76f79-wzgcr" podStartSLOduration=3.27643946 podStartE2EDuration="3.27643946s" podCreationTimestamp="2026-03-11 13:25:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:25:55.275635418 +0000 UTC m=+202.957537162" watchObservedRunningTime="2026-03-11 13:25:55.27643946 +0000 UTC m=+202.958341174" Mar 11 13:25:56 crc kubenswrapper[4923]: I0311 13:25:56.265567 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-75d8cbb545-xmbff" event={"ID":"57ac6c22-6124-437d-bbe7-59c376badfc4","Type":"ContainerStarted","Data":"b5f80cc6c35b203d712fae7c184efc06a938b57d43c63b518fd77157de493df1"} Mar 11 13:25:56 crc kubenswrapper[4923]: I0311 13:25:56.265632 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-75d8cbb545-xmbff" Mar 11 13:25:56 crc kubenswrapper[4923]: I0311 13:25:56.272641 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-75d8cbb545-xmbff" Mar 11 13:25:56 crc kubenswrapper[4923]: I0311 13:25:56.294601 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-75d8cbb545-xmbff" podStartSLOduration=4.29458292 podStartE2EDuration="4.29458292s" podCreationTimestamp="2026-03-11 13:25:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:25:56.290033304 +0000 UTC m=+203.971935048" watchObservedRunningTime="2026-03-11 13:25:56.29458292 +0000 UTC m=+203.976484624" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.683086 4923 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.684939 4923 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.685198 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.685607 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://b865dff977394535b945fbacba9ca628914f266d5e6a4bf1281a6771d884f622" gracePeriod=15 Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.685663 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://c84697dbf2629c6bfcdf82a7ff158126b4c1ec298d635619eff5fbe84af05118" gracePeriod=15 Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.685723 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://8310025c89cd82e434fa820785f24adeb870dc4c436e02258252b3a8493dc6ca" gracePeriod=15 Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.685755 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://b1e1d3884306e5cb7b7d8e416ba17caa145f530c0f6d7b35a81ef10f4a79fadb" gracePeriod=15 Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.686069 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://2a2537d234646cc42128c4e97220f789c9bcbe25c060c0f488aa820e3f611848" gracePeriod=15 Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.687039 4923 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 11 13:25:58 crc kubenswrapper[4923]: E0311 13:25:58.691682 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.691740 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 11 13:25:58 crc kubenswrapper[4923]: E0311 13:25:58.692091 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.692118 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 11 13:25:58 crc kubenswrapper[4923]: E0311 13:25:58.692141 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.692158 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 11 13:25:58 crc kubenswrapper[4923]: E0311 13:25:58.692181 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.692232 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 11 13:25:58 crc kubenswrapper[4923]: E0311 13:25:58.692270 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.692286 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 11 13:25:58 crc kubenswrapper[4923]: E0311 13:25:58.692322 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.692338 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 11 13:25:58 crc kubenswrapper[4923]: E0311 13:25:58.695588 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.695614 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 11 13:25:58 crc kubenswrapper[4923]: E0311 13:25:58.695639 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.695655 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.696633 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.696681 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.696711 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.696741 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.696762 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.696801 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.696866 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.696904 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 11 13:25:58 crc kubenswrapper[4923]: E0311 13:25:58.700480 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.700533 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 11 13:25:58 crc kubenswrapper[4923]: E0311 13:25:58.700574 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.700592 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.701140 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 11 13:25:58 crc kubenswrapper[4923]: E0311 13:25:58.775225 4923 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.111:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.784585 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.784648 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.784692 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.784727 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.784766 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.784797 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.784835 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.784866 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.885613 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.885898 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.886057 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.886238 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.886396 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.886539 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.886626 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.886317 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.885831 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.886471 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.886153 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.885977 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.886850 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.886650 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.887044 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:25:58 crc kubenswrapper[4923]: I0311 13:25:58.887096 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:25:59 crc kubenswrapper[4923]: I0311 13:25:59.077191 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 11 13:25:59 crc kubenswrapper[4923]: E0311 13:25:59.103557 4923 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.111:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189bcc58935d19c5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:25:59.101962693 +0000 UTC m=+206.783864447,LastTimestamp:2026-03-11 13:25:59.101962693 +0000 UTC m=+206.783864447,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:25:59 crc kubenswrapper[4923]: I0311 13:25:59.290759 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"851feb81d6f20afbb5871cc3b0749ef065db06d3dc3224e8d0c9ccfbf1f0802e"} Mar 11 13:25:59 crc kubenswrapper[4923]: I0311 13:25:59.295151 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 11 13:25:59 crc kubenswrapper[4923]: I0311 13:25:59.297153 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 11 13:25:59 crc kubenswrapper[4923]: I0311 13:25:59.298005 4923 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c84697dbf2629c6bfcdf82a7ff158126b4c1ec298d635619eff5fbe84af05118" exitCode=0 Mar 11 13:25:59 crc kubenswrapper[4923]: I0311 13:25:59.298049 4923 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8310025c89cd82e434fa820785f24adeb870dc4c436e02258252b3a8493dc6ca" exitCode=0 Mar 11 13:25:59 crc kubenswrapper[4923]: I0311 13:25:59.298065 4923 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2a2537d234646cc42128c4e97220f789c9bcbe25c060c0f488aa820e3f611848" exitCode=0 Mar 11 13:25:59 crc kubenswrapper[4923]: I0311 13:25:59.298080 4923 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b1e1d3884306e5cb7b7d8e416ba17caa145f530c0f6d7b35a81ef10f4a79fadb" exitCode=2 Mar 11 13:25:59 crc kubenswrapper[4923]: I0311 13:25:59.298164 4923 scope.go:117] "RemoveContainer" containerID="9a9f78b1ff5dfa00fd3d80bafc5c131e5fd12de6d0556229087316aad81df459" Mar 11 13:25:59 crc kubenswrapper[4923]: I0311 13:25:59.300405 4923 generic.go:334] "Generic (PLEG): container finished" podID="331aaa7c-9dd6-4f2c-b3d8-00febd6448b8" containerID="32f264786f21fe350bcdecb293da2b7dcf1e5cecba0e4ac8023fd116eb0f5a69" exitCode=0 Mar 11 13:25:59 crc kubenswrapper[4923]: I0311 13:25:59.300452 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"331aaa7c-9dd6-4f2c-b3d8-00febd6448b8","Type":"ContainerDied","Data":"32f264786f21fe350bcdecb293da2b7dcf1e5cecba0e4ac8023fd116eb0f5a69"} Mar 11 13:25:59 crc kubenswrapper[4923]: I0311 13:25:59.301443 4923 status_manager.go:851] "Failed to get status for pod" podUID="331aaa7c-9dd6-4f2c-b3d8-00febd6448b8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.111:6443: connect: connection refused" Mar 11 13:26:00 crc kubenswrapper[4923]: I0311 13:26:00.310558 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 11 13:26:00 crc kubenswrapper[4923]: I0311 13:26:00.313671 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"8839419742b703ea5a1d3db097f4d336d55ae25bb99d9212dffa808eec14601f"} Mar 11 13:26:00 crc kubenswrapper[4923]: E0311 13:26:00.314460 4923 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.111:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 11 13:26:00 crc kubenswrapper[4923]: I0311 13:26:00.314473 4923 status_manager.go:851] "Failed to get status for pod" podUID="331aaa7c-9dd6-4f2c-b3d8-00febd6448b8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.111:6443: connect: connection refused" Mar 11 13:26:00 crc kubenswrapper[4923]: I0311 13:26:00.712228 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 11 13:26:00 crc kubenswrapper[4923]: I0311 13:26:00.713709 4923 status_manager.go:851] "Failed to get status for pod" podUID="331aaa7c-9dd6-4f2c-b3d8-00febd6448b8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.111:6443: connect: connection refused" Mar 11 13:26:00 crc kubenswrapper[4923]: I0311 13:26:00.811328 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/331aaa7c-9dd6-4f2c-b3d8-00febd6448b8-kube-api-access\") pod \"331aaa7c-9dd6-4f2c-b3d8-00febd6448b8\" (UID: \"331aaa7c-9dd6-4f2c-b3d8-00febd6448b8\") " Mar 11 13:26:00 crc kubenswrapper[4923]: I0311 13:26:00.815578 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/331aaa7c-9dd6-4f2c-b3d8-00febd6448b8-kubelet-dir\") pod \"331aaa7c-9dd6-4f2c-b3d8-00febd6448b8\" (UID: \"331aaa7c-9dd6-4f2c-b3d8-00febd6448b8\") " Mar 11 13:26:00 crc kubenswrapper[4923]: I0311 13:26:00.815636 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/331aaa7c-9dd6-4f2c-b3d8-00febd6448b8-var-lock\") pod \"331aaa7c-9dd6-4f2c-b3d8-00febd6448b8\" (UID: \"331aaa7c-9dd6-4f2c-b3d8-00febd6448b8\") " Mar 11 13:26:00 crc kubenswrapper[4923]: I0311 13:26:00.816261 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/331aaa7c-9dd6-4f2c-b3d8-00febd6448b8-var-lock" (OuterVolumeSpecName: "var-lock") pod "331aaa7c-9dd6-4f2c-b3d8-00febd6448b8" (UID: "331aaa7c-9dd6-4f2c-b3d8-00febd6448b8"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:26:00 crc kubenswrapper[4923]: I0311 13:26:00.816310 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/331aaa7c-9dd6-4f2c-b3d8-00febd6448b8-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "331aaa7c-9dd6-4f2c-b3d8-00febd6448b8" (UID: "331aaa7c-9dd6-4f2c-b3d8-00febd6448b8"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:26:00 crc kubenswrapper[4923]: I0311 13:26:00.836765 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/331aaa7c-9dd6-4f2c-b3d8-00febd6448b8-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "331aaa7c-9dd6-4f2c-b3d8-00febd6448b8" (UID: "331aaa7c-9dd6-4f2c-b3d8-00febd6448b8"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:26:00 crc kubenswrapper[4923]: I0311 13:26:00.918119 4923 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/331aaa7c-9dd6-4f2c-b3d8-00febd6448b8-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 11 13:26:00 crc kubenswrapper[4923]: I0311 13:26:00.918153 4923 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/331aaa7c-9dd6-4f2c-b3d8-00febd6448b8-var-lock\") on node \"crc\" DevicePath \"\"" Mar 11 13:26:00 crc kubenswrapper[4923]: I0311 13:26:00.918165 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/331aaa7c-9dd6-4f2c-b3d8-00febd6448b8-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.107089 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.108525 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.109191 4923 status_manager.go:851] "Failed to get status for pod" podUID="331aaa7c-9dd6-4f2c-b3d8-00febd6448b8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.111:6443: connect: connection refused" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.109657 4923 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.111:6443: connect: connection refused" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.223264 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.223459 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.223576 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.223676 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.223810 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.223722 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.224212 4923 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.224247 4923 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.224266 4923 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.326915 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.327944 4923 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b865dff977394535b945fbacba9ca628914f266d5e6a4bf1281a6771d884f622" exitCode=0 Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.328087 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.328203 4923 scope.go:117] "RemoveContainer" containerID="c84697dbf2629c6bfcdf82a7ff158126b4c1ec298d635619eff5fbe84af05118" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.330858 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"331aaa7c-9dd6-4f2c-b3d8-00febd6448b8","Type":"ContainerDied","Data":"4cef773813c29b7c700536a9913586524b812d852e9058e3c6ff5a9c38e3aed5"} Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.330905 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4cef773813c29b7c700536a9913586524b812d852e9058e3c6ff5a9c38e3aed5" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.331229 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 11 13:26:01 crc kubenswrapper[4923]: E0311 13:26:01.331989 4923 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.111:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.344303 4923 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.111:6443: connect: connection refused" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.344835 4923 status_manager.go:851] "Failed to get status for pod" podUID="331aaa7c-9dd6-4f2c-b3d8-00febd6448b8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.111:6443: connect: connection refused" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.345273 4923 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.111:6443: connect: connection refused" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.345682 4923 status_manager.go:851] "Failed to get status for pod" podUID="331aaa7c-9dd6-4f2c-b3d8-00febd6448b8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.111:6443: connect: connection refused" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.355093 4923 scope.go:117] "RemoveContainer" containerID="8310025c89cd82e434fa820785f24adeb870dc4c436e02258252b3a8493dc6ca" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.382210 4923 scope.go:117] "RemoveContainer" containerID="2a2537d234646cc42128c4e97220f789c9bcbe25c060c0f488aa820e3f611848" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.402150 4923 scope.go:117] "RemoveContainer" containerID="b1e1d3884306e5cb7b7d8e416ba17caa145f530c0f6d7b35a81ef10f4a79fadb" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.428295 4923 scope.go:117] "RemoveContainer" containerID="b865dff977394535b945fbacba9ca628914f266d5e6a4bf1281a6771d884f622" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.446041 4923 scope.go:117] "RemoveContainer" containerID="6e1f00d5974fffbe97d7d9d7dc2c8eab8a73b272ee66d9633c90ca97bd23ded1" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.478922 4923 scope.go:117] "RemoveContainer" containerID="c84697dbf2629c6bfcdf82a7ff158126b4c1ec298d635619eff5fbe84af05118" Mar 11 13:26:01 crc kubenswrapper[4923]: E0311 13:26:01.479439 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c84697dbf2629c6bfcdf82a7ff158126b4c1ec298d635619eff5fbe84af05118\": container with ID starting with c84697dbf2629c6bfcdf82a7ff158126b4c1ec298d635619eff5fbe84af05118 not found: ID does not exist" containerID="c84697dbf2629c6bfcdf82a7ff158126b4c1ec298d635619eff5fbe84af05118" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.479483 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c84697dbf2629c6bfcdf82a7ff158126b4c1ec298d635619eff5fbe84af05118"} err="failed to get container status \"c84697dbf2629c6bfcdf82a7ff158126b4c1ec298d635619eff5fbe84af05118\": rpc error: code = NotFound desc = could not find container \"c84697dbf2629c6bfcdf82a7ff158126b4c1ec298d635619eff5fbe84af05118\": container with ID starting with c84697dbf2629c6bfcdf82a7ff158126b4c1ec298d635619eff5fbe84af05118 not found: ID does not exist" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.479536 4923 scope.go:117] "RemoveContainer" containerID="8310025c89cd82e434fa820785f24adeb870dc4c436e02258252b3a8493dc6ca" Mar 11 13:26:01 crc kubenswrapper[4923]: E0311 13:26:01.480238 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8310025c89cd82e434fa820785f24adeb870dc4c436e02258252b3a8493dc6ca\": container with ID starting with 8310025c89cd82e434fa820785f24adeb870dc4c436e02258252b3a8493dc6ca not found: ID does not exist" containerID="8310025c89cd82e434fa820785f24adeb870dc4c436e02258252b3a8493dc6ca" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.480373 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8310025c89cd82e434fa820785f24adeb870dc4c436e02258252b3a8493dc6ca"} err="failed to get container status \"8310025c89cd82e434fa820785f24adeb870dc4c436e02258252b3a8493dc6ca\": rpc error: code = NotFound desc = could not find container \"8310025c89cd82e434fa820785f24adeb870dc4c436e02258252b3a8493dc6ca\": container with ID starting with 8310025c89cd82e434fa820785f24adeb870dc4c436e02258252b3a8493dc6ca not found: ID does not exist" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.480434 4923 scope.go:117] "RemoveContainer" containerID="2a2537d234646cc42128c4e97220f789c9bcbe25c060c0f488aa820e3f611848" Mar 11 13:26:01 crc kubenswrapper[4923]: E0311 13:26:01.480923 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a2537d234646cc42128c4e97220f789c9bcbe25c060c0f488aa820e3f611848\": container with ID starting with 2a2537d234646cc42128c4e97220f789c9bcbe25c060c0f488aa820e3f611848 not found: ID does not exist" containerID="2a2537d234646cc42128c4e97220f789c9bcbe25c060c0f488aa820e3f611848" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.480971 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a2537d234646cc42128c4e97220f789c9bcbe25c060c0f488aa820e3f611848"} err="failed to get container status \"2a2537d234646cc42128c4e97220f789c9bcbe25c060c0f488aa820e3f611848\": rpc error: code = NotFound desc = could not find container \"2a2537d234646cc42128c4e97220f789c9bcbe25c060c0f488aa820e3f611848\": container with ID starting with 2a2537d234646cc42128c4e97220f789c9bcbe25c060c0f488aa820e3f611848 not found: ID does not exist" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.480998 4923 scope.go:117] "RemoveContainer" containerID="b1e1d3884306e5cb7b7d8e416ba17caa145f530c0f6d7b35a81ef10f4a79fadb" Mar 11 13:26:01 crc kubenswrapper[4923]: E0311 13:26:01.481334 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1e1d3884306e5cb7b7d8e416ba17caa145f530c0f6d7b35a81ef10f4a79fadb\": container with ID starting with b1e1d3884306e5cb7b7d8e416ba17caa145f530c0f6d7b35a81ef10f4a79fadb not found: ID does not exist" containerID="b1e1d3884306e5cb7b7d8e416ba17caa145f530c0f6d7b35a81ef10f4a79fadb" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.481396 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1e1d3884306e5cb7b7d8e416ba17caa145f530c0f6d7b35a81ef10f4a79fadb"} err="failed to get container status \"b1e1d3884306e5cb7b7d8e416ba17caa145f530c0f6d7b35a81ef10f4a79fadb\": rpc error: code = NotFound desc = could not find container \"b1e1d3884306e5cb7b7d8e416ba17caa145f530c0f6d7b35a81ef10f4a79fadb\": container with ID starting with b1e1d3884306e5cb7b7d8e416ba17caa145f530c0f6d7b35a81ef10f4a79fadb not found: ID does not exist" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.481428 4923 scope.go:117] "RemoveContainer" containerID="b865dff977394535b945fbacba9ca628914f266d5e6a4bf1281a6771d884f622" Mar 11 13:26:01 crc kubenswrapper[4923]: E0311 13:26:01.481824 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b865dff977394535b945fbacba9ca628914f266d5e6a4bf1281a6771d884f622\": container with ID starting with b865dff977394535b945fbacba9ca628914f266d5e6a4bf1281a6771d884f622 not found: ID does not exist" containerID="b865dff977394535b945fbacba9ca628914f266d5e6a4bf1281a6771d884f622" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.481853 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b865dff977394535b945fbacba9ca628914f266d5e6a4bf1281a6771d884f622"} err="failed to get container status \"b865dff977394535b945fbacba9ca628914f266d5e6a4bf1281a6771d884f622\": rpc error: code = NotFound desc = could not find container \"b865dff977394535b945fbacba9ca628914f266d5e6a4bf1281a6771d884f622\": container with ID starting with b865dff977394535b945fbacba9ca628914f266d5e6a4bf1281a6771d884f622 not found: ID does not exist" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.481873 4923 scope.go:117] "RemoveContainer" containerID="6e1f00d5974fffbe97d7d9d7dc2c8eab8a73b272ee66d9633c90ca97bd23ded1" Mar 11 13:26:01 crc kubenswrapper[4923]: E0311 13:26:01.482159 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e1f00d5974fffbe97d7d9d7dc2c8eab8a73b272ee66d9633c90ca97bd23ded1\": container with ID starting with 6e1f00d5974fffbe97d7d9d7dc2c8eab8a73b272ee66d9633c90ca97bd23ded1 not found: ID does not exist" containerID="6e1f00d5974fffbe97d7d9d7dc2c8eab8a73b272ee66d9633c90ca97bd23ded1" Mar 11 13:26:01 crc kubenswrapper[4923]: I0311 13:26:01.482183 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e1f00d5974fffbe97d7d9d7dc2c8eab8a73b272ee66d9633c90ca97bd23ded1"} err="failed to get container status \"6e1f00d5974fffbe97d7d9d7dc2c8eab8a73b272ee66d9633c90ca97bd23ded1\": rpc error: code = NotFound desc = could not find container \"6e1f00d5974fffbe97d7d9d7dc2c8eab8a73b272ee66d9633c90ca97bd23ded1\": container with ID starting with 6e1f00d5974fffbe97d7d9d7dc2c8eab8a73b272ee66d9633c90ca97bd23ded1 not found: ID does not exist" Mar 11 13:26:03 crc kubenswrapper[4923]: I0311 13:26:03.025304 4923 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.111:6443: connect: connection refused" Mar 11 13:26:03 crc kubenswrapper[4923]: I0311 13:26:03.027668 4923 status_manager.go:851] "Failed to get status for pod" podUID="331aaa7c-9dd6-4f2c-b3d8-00febd6448b8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.111:6443: connect: connection refused" Mar 11 13:26:03 crc kubenswrapper[4923]: I0311 13:26:03.036052 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 11 13:26:06 crc kubenswrapper[4923]: E0311 13:26:06.288084 4923 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.111:6443: connect: connection refused" Mar 11 13:26:06 crc kubenswrapper[4923]: E0311 13:26:06.288953 4923 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.111:6443: connect: connection refused" Mar 11 13:26:06 crc kubenswrapper[4923]: E0311 13:26:06.289632 4923 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.111:6443: connect: connection refused" Mar 11 13:26:06 crc kubenswrapper[4923]: E0311 13:26:06.290234 4923 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.111:6443: connect: connection refused" Mar 11 13:26:06 crc kubenswrapper[4923]: E0311 13:26:06.290786 4923 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.111:6443: connect: connection refused" Mar 11 13:26:06 crc kubenswrapper[4923]: I0311 13:26:06.290864 4923 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 11 13:26:06 crc kubenswrapper[4923]: E0311 13:26:06.291412 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.111:6443: connect: connection refused" interval="200ms" Mar 11 13:26:06 crc kubenswrapper[4923]: E0311 13:26:06.492730 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.111:6443: connect: connection refused" interval="400ms" Mar 11 13:26:06 crc kubenswrapper[4923]: E0311 13:26:06.893968 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.111:6443: connect: connection refused" interval="800ms" Mar 11 13:26:07 crc kubenswrapper[4923]: E0311 13:26:07.161155 4923 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.111:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189bcc58935d19c5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-11 13:25:59.101962693 +0000 UTC m=+206.783864447,LastTimestamp:2026-03-11 13:25:59.101962693 +0000 UTC m=+206.783864447,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 11 13:26:07 crc kubenswrapper[4923]: E0311 13:26:07.695575 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.111:6443: connect: connection refused" interval="1.6s" Mar 11 13:26:08 crc kubenswrapper[4923]: E0311 13:26:08.103256 4923 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.111:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" volumeName="registry-storage" Mar 11 13:26:09 crc kubenswrapper[4923]: E0311 13:26:09.296851 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.111:6443: connect: connection refused" interval="3.2s" Mar 11 13:26:10 crc kubenswrapper[4923]: I0311 13:26:10.021456 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:26:10 crc kubenswrapper[4923]: I0311 13:26:10.022312 4923 status_manager.go:851] "Failed to get status for pod" podUID="331aaa7c-9dd6-4f2c-b3d8-00febd6448b8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.111:6443: connect: connection refused" Mar 11 13:26:10 crc kubenswrapper[4923]: I0311 13:26:10.050261 4923 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d2b0d8e-7f7b-4678-a3b2-ef1831fb1574" Mar 11 13:26:10 crc kubenswrapper[4923]: I0311 13:26:10.050327 4923 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d2b0d8e-7f7b-4678-a3b2-ef1831fb1574" Mar 11 13:26:10 crc kubenswrapper[4923]: E0311 13:26:10.050998 4923 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.111:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:26:10 crc kubenswrapper[4923]: I0311 13:26:10.051737 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:26:10 crc kubenswrapper[4923]: W0311 13:26:10.086949 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-ce4a1e47cd7a501e2940fad39e9fb63b8bc5014f9be4c6c570ed20a26a09c9a8 WatchSource:0}: Error finding container ce4a1e47cd7a501e2940fad39e9fb63b8bc5014f9be4c6c570ed20a26a09c9a8: Status 404 returned error can't find the container with id ce4a1e47cd7a501e2940fad39e9fb63b8bc5014f9be4c6c570ed20a26a09c9a8 Mar 11 13:26:10 crc kubenswrapper[4923]: I0311 13:26:10.393661 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ce4a1e47cd7a501e2940fad39e9fb63b8bc5014f9be4c6c570ed20a26a09c9a8"} Mar 11 13:26:11 crc kubenswrapper[4923]: I0311 13:26:11.403280 4923 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="936b6984da95731c611530f89c8c36c786f5f5d72c3242257ae57e4d05e47a96" exitCode=0 Mar 11 13:26:11 crc kubenswrapper[4923]: I0311 13:26:11.403524 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"936b6984da95731c611530f89c8c36c786f5f5d72c3242257ae57e4d05e47a96"} Mar 11 13:26:11 crc kubenswrapper[4923]: I0311 13:26:11.403903 4923 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d2b0d8e-7f7b-4678-a3b2-ef1831fb1574" Mar 11 13:26:11 crc kubenswrapper[4923]: I0311 13:26:11.403942 4923 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d2b0d8e-7f7b-4678-a3b2-ef1831fb1574" Mar 11 13:26:11 crc kubenswrapper[4923]: I0311 13:26:11.404334 4923 status_manager.go:851] "Failed to get status for pod" podUID="331aaa7c-9dd6-4f2c-b3d8-00febd6448b8" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.111:6443: connect: connection refused" Mar 11 13:26:11 crc kubenswrapper[4923]: E0311 13:26:11.404420 4923 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.111:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:26:12 crc kubenswrapper[4923]: I0311 13:26:12.417951 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c6910c0a4235878c51a6c6a6332da1c4061e06068995695c81ed704ac1fc3f8f"} Mar 11 13:26:12 crc kubenswrapper[4923]: I0311 13:26:12.418007 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a5b6c894a28600e547370e958451f542fd88287ff587b86a5508930723694efe"} Mar 11 13:26:12 crc kubenswrapper[4923]: I0311 13:26:12.418026 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"44c86fc7bb4fe4dcec4339503e66679d342607d2e4090309dcf3741c9c2218e6"} Mar 11 13:26:13 crc kubenswrapper[4923]: I0311 13:26:13.426363 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"422335a14e2d15141abbbb6a92fed104c4c07154ff63092fdf4203b6eb147552"} Mar 11 13:26:13 crc kubenswrapper[4923]: I0311 13:26:13.426614 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"813ee3928f393f63410edae43d29920a4ee5cd7e093df2cb2bc08530bb8611b0"} Mar 11 13:26:13 crc kubenswrapper[4923]: I0311 13:26:13.426804 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:26:13 crc kubenswrapper[4923]: I0311 13:26:13.427022 4923 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d2b0d8e-7f7b-4678-a3b2-ef1831fb1574" Mar 11 13:26:13 crc kubenswrapper[4923]: I0311 13:26:13.427066 4923 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d2b0d8e-7f7b-4678-a3b2-ef1831fb1574" Mar 11 13:26:13 crc kubenswrapper[4923]: I0311 13:26:13.429283 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 11 13:26:13 crc kubenswrapper[4923]: I0311 13:26:13.429854 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 11 13:26:13 crc kubenswrapper[4923]: I0311 13:26:13.429923 4923 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="4b29200103f0d31fe398058b057f9843b9eec00edc4d3ed2da9f794618ca50c8" exitCode=1 Mar 11 13:26:13 crc kubenswrapper[4923]: I0311 13:26:13.429955 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"4b29200103f0d31fe398058b057f9843b9eec00edc4d3ed2da9f794618ca50c8"} Mar 11 13:26:13 crc kubenswrapper[4923]: I0311 13:26:13.430517 4923 scope.go:117] "RemoveContainer" containerID="4b29200103f0d31fe398058b057f9843b9eec00edc4d3ed2da9f794618ca50c8" Mar 11 13:26:14 crc kubenswrapper[4923]: I0311 13:26:14.252725 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:26:14 crc kubenswrapper[4923]: I0311 13:26:14.443363 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 11 13:26:14 crc kubenswrapper[4923]: I0311 13:26:14.444430 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 11 13:26:14 crc kubenswrapper[4923]: I0311 13:26:14.444536 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"39b26c9ebd2f3e7c6294220bf68042bc1f36807ba7f85ddeffb8d9fc1898e56b"} Mar 11 13:26:15 crc kubenswrapper[4923]: I0311 13:26:15.052456 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:26:15 crc kubenswrapper[4923]: I0311 13:26:15.052735 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:26:15 crc kubenswrapper[4923]: I0311 13:26:15.067694 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:26:18 crc kubenswrapper[4923]: I0311 13:26:18.445362 4923 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:26:18 crc kubenswrapper[4923]: I0311 13:26:18.446990 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6d2b0d8e-7f7b-4678-a3b2-ef1831fb1574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:26:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:26:11Z\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:26:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-cert-syncer kube-apiserver-cert-regeneration-controller kube-apiserver-insecure-readyz kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-11T13:26:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-cert-syncer kube-apiserver-cert-regeneration-controller kube-apiserver-insecure-readyz kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}}}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://936b6984da95731c611530f89c8c36c786f5f5d72c3242257ae57e4d05e47a96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://936b6984da95731c611530f89c8c36c786f5f5d72c3242257ae57e4d05e47a96\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-11T13:26:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-11T13:26:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Pending\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": pods \"kube-apiserver-crc\" not found" Mar 11 13:26:18 crc kubenswrapper[4923]: I0311 13:26:18.470639 4923 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d2b0d8e-7f7b-4678-a3b2-ef1831fb1574" Mar 11 13:26:18 crc kubenswrapper[4923]: I0311 13:26:18.470674 4923 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d2b0d8e-7f7b-4678-a3b2-ef1831fb1574" Mar 11 13:26:18 crc kubenswrapper[4923]: I0311 13:26:18.475659 4923 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="ca44a61e-9cde-443a-8977-6541c0c3c06e" Mar 11 13:26:18 crc kubenswrapper[4923]: I0311 13:26:18.476910 4923 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://44c86fc7bb4fe4dcec4339503e66679d342607d2e4090309dcf3741c9c2218e6" Mar 11 13:26:18 crc kubenswrapper[4923]: I0311 13:26:18.476940 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:26:19 crc kubenswrapper[4923]: I0311 13:26:19.476569 4923 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d2b0d8e-7f7b-4678-a3b2-ef1831fb1574" Mar 11 13:26:19 crc kubenswrapper[4923]: I0311 13:26:19.476956 4923 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="6d2b0d8e-7f7b-4678-a3b2-ef1831fb1574" Mar 11 13:26:19 crc kubenswrapper[4923]: I0311 13:26:19.479729 4923 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="ca44a61e-9cde-443a-8977-6541c0c3c06e" Mar 11 13:26:21 crc kubenswrapper[4923]: I0311 13:26:21.683949 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:26:24 crc kubenswrapper[4923]: I0311 13:26:24.252717 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:26:24 crc kubenswrapper[4923]: I0311 13:26:24.260870 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:26:24 crc kubenswrapper[4923]: I0311 13:26:24.523601 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 11 13:26:29 crc kubenswrapper[4923]: I0311 13:26:29.008762 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 11 13:26:29 crc kubenswrapper[4923]: I0311 13:26:29.464500 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 11 13:26:29 crc kubenswrapper[4923]: I0311 13:26:29.533609 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 11 13:26:29 crc kubenswrapper[4923]: I0311 13:26:29.573845 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 11 13:26:30 crc kubenswrapper[4923]: I0311 13:26:30.005452 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 11 13:26:30 crc kubenswrapper[4923]: I0311 13:26:30.501898 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 11 13:26:30 crc kubenswrapper[4923]: I0311 13:26:30.528269 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 11 13:26:30 crc kubenswrapper[4923]: I0311 13:26:30.922576 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 11 13:26:30 crc kubenswrapper[4923]: I0311 13:26:30.933328 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 11 13:26:31 crc kubenswrapper[4923]: I0311 13:26:31.028403 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 11 13:26:31 crc kubenswrapper[4923]: I0311 13:26:31.154088 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 11 13:26:31 crc kubenswrapper[4923]: I0311 13:26:31.280463 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 11 13:26:31 crc kubenswrapper[4923]: I0311 13:26:31.397408 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 11 13:26:31 crc kubenswrapper[4923]: I0311 13:26:31.625057 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 11 13:26:31 crc kubenswrapper[4923]: I0311 13:26:31.924293 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 11 13:26:31 crc kubenswrapper[4923]: I0311 13:26:31.936716 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 11 13:26:31 crc kubenswrapper[4923]: I0311 13:26:31.964620 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 11 13:26:31 crc kubenswrapper[4923]: I0311 13:26:31.969069 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 11 13:26:31 crc kubenswrapper[4923]: I0311 13:26:31.973874 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 11 13:26:32 crc kubenswrapper[4923]: I0311 13:26:32.101452 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 11 13:26:32 crc kubenswrapper[4923]: I0311 13:26:32.142474 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 11 13:26:32 crc kubenswrapper[4923]: I0311 13:26:32.144786 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 11 13:26:32 crc kubenswrapper[4923]: I0311 13:26:32.198091 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 11 13:26:32 crc kubenswrapper[4923]: I0311 13:26:32.261909 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 11 13:26:32 crc kubenswrapper[4923]: I0311 13:26:32.419365 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 11 13:26:32 crc kubenswrapper[4923]: I0311 13:26:32.443038 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 11 13:26:32 crc kubenswrapper[4923]: I0311 13:26:32.470008 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 11 13:26:32 crc kubenswrapper[4923]: I0311 13:26:32.585238 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 11 13:26:32 crc kubenswrapper[4923]: I0311 13:26:32.641867 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 11 13:26:32 crc kubenswrapper[4923]: I0311 13:26:32.842085 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 11 13:26:32 crc kubenswrapper[4923]: I0311 13:26:32.955333 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 11 13:26:32 crc kubenswrapper[4923]: I0311 13:26:32.968149 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 11 13:26:33 crc kubenswrapper[4923]: I0311 13:26:33.008074 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 11 13:26:33 crc kubenswrapper[4923]: I0311 13:26:33.063469 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 11 13:26:33 crc kubenswrapper[4923]: I0311 13:26:33.096025 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 11 13:26:33 crc kubenswrapper[4923]: I0311 13:26:33.159395 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 11 13:26:33 crc kubenswrapper[4923]: I0311 13:26:33.160420 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 11 13:26:33 crc kubenswrapper[4923]: I0311 13:26:33.181305 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 11 13:26:33 crc kubenswrapper[4923]: I0311 13:26:33.209631 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 11 13:26:33 crc kubenswrapper[4923]: I0311 13:26:33.265387 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 11 13:26:33 crc kubenswrapper[4923]: I0311 13:26:33.282873 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 11 13:26:33 crc kubenswrapper[4923]: I0311 13:26:33.336972 4923 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 11 13:26:33 crc kubenswrapper[4923]: I0311 13:26:33.346419 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 11 13:26:33 crc kubenswrapper[4923]: I0311 13:26:33.479772 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 11 13:26:33 crc kubenswrapper[4923]: I0311 13:26:33.524564 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 11 13:26:33 crc kubenswrapper[4923]: I0311 13:26:33.604072 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 11 13:26:33 crc kubenswrapper[4923]: I0311 13:26:33.689588 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 11 13:26:33 crc kubenswrapper[4923]: I0311 13:26:33.807177 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 11 13:26:33 crc kubenswrapper[4923]: I0311 13:26:33.864008 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 11 13:26:33 crc kubenswrapper[4923]: I0311 13:26:33.879154 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 11 13:26:33 crc kubenswrapper[4923]: I0311 13:26:33.930639 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 11 13:26:34 crc kubenswrapper[4923]: I0311 13:26:34.169245 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 11 13:26:34 crc kubenswrapper[4923]: I0311 13:26:34.296986 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 11 13:26:34 crc kubenswrapper[4923]: I0311 13:26:34.461484 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 11 13:26:34 crc kubenswrapper[4923]: I0311 13:26:34.508828 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 11 13:26:34 crc kubenswrapper[4923]: I0311 13:26:34.578942 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 11 13:26:34 crc kubenswrapper[4923]: I0311 13:26:34.637627 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 11 13:26:34 crc kubenswrapper[4923]: I0311 13:26:34.680752 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 11 13:26:34 crc kubenswrapper[4923]: I0311 13:26:34.683558 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 11 13:26:34 crc kubenswrapper[4923]: I0311 13:26:34.697905 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 11 13:26:34 crc kubenswrapper[4923]: I0311 13:26:34.757912 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 11 13:26:34 crc kubenswrapper[4923]: I0311 13:26:34.784453 4923 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 11 13:26:34 crc kubenswrapper[4923]: I0311 13:26:34.807210 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 11 13:26:34 crc kubenswrapper[4923]: I0311 13:26:34.855288 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 11 13:26:34 crc kubenswrapper[4923]: I0311 13:26:34.858283 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 11 13:26:34 crc kubenswrapper[4923]: I0311 13:26:34.872731 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 11 13:26:34 crc kubenswrapper[4923]: I0311 13:26:34.929009 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 11 13:26:34 crc kubenswrapper[4923]: I0311 13:26:34.941062 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 11 13:26:34 crc kubenswrapper[4923]: I0311 13:26:34.963309 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 11 13:26:34 crc kubenswrapper[4923]: I0311 13:26:34.980866 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 11 13:26:35 crc kubenswrapper[4923]: I0311 13:26:35.018908 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 11 13:26:35 crc kubenswrapper[4923]: I0311 13:26:35.091575 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 11 13:26:35 crc kubenswrapper[4923]: I0311 13:26:35.124489 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 11 13:26:35 crc kubenswrapper[4923]: I0311 13:26:35.255406 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 11 13:26:35 crc kubenswrapper[4923]: I0311 13:26:35.548911 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 11 13:26:35 crc kubenswrapper[4923]: I0311 13:26:35.552800 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 11 13:26:35 crc kubenswrapper[4923]: I0311 13:26:35.563818 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 11 13:26:35 crc kubenswrapper[4923]: I0311 13:26:35.629927 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 11 13:26:35 crc kubenswrapper[4923]: I0311 13:26:35.642837 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 11 13:26:35 crc kubenswrapper[4923]: I0311 13:26:35.682049 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 11 13:26:35 crc kubenswrapper[4923]: I0311 13:26:35.697434 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 11 13:26:35 crc kubenswrapper[4923]: I0311 13:26:35.901293 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 11 13:26:35 crc kubenswrapper[4923]: I0311 13:26:35.931415 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 11 13:26:35 crc kubenswrapper[4923]: I0311 13:26:35.941476 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 11 13:26:35 crc kubenswrapper[4923]: I0311 13:26:35.959758 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 11 13:26:35 crc kubenswrapper[4923]: I0311 13:26:35.972781 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 11 13:26:35 crc kubenswrapper[4923]: I0311 13:26:35.984817 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 11 13:26:36 crc kubenswrapper[4923]: I0311 13:26:36.048016 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 11 13:26:36 crc kubenswrapper[4923]: I0311 13:26:36.099399 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 11 13:26:36 crc kubenswrapper[4923]: I0311 13:26:36.103989 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 11 13:26:36 crc kubenswrapper[4923]: I0311 13:26:36.113307 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 11 13:26:36 crc kubenswrapper[4923]: I0311 13:26:36.186583 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 11 13:26:36 crc kubenswrapper[4923]: I0311 13:26:36.292186 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 11 13:26:36 crc kubenswrapper[4923]: I0311 13:26:36.304316 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 11 13:26:36 crc kubenswrapper[4923]: I0311 13:26:36.358260 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 11 13:26:36 crc kubenswrapper[4923]: I0311 13:26:36.469131 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 11 13:26:36 crc kubenswrapper[4923]: I0311 13:26:36.470244 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 11 13:26:36 crc kubenswrapper[4923]: I0311 13:26:36.506960 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 11 13:26:36 crc kubenswrapper[4923]: I0311 13:26:36.607519 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 11 13:26:36 crc kubenswrapper[4923]: I0311 13:26:36.709441 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 11 13:26:36 crc kubenswrapper[4923]: I0311 13:26:36.780718 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 11 13:26:36 crc kubenswrapper[4923]: I0311 13:26:36.801526 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 11 13:26:36 crc kubenswrapper[4923]: I0311 13:26:36.814634 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 11 13:26:36 crc kubenswrapper[4923]: I0311 13:26:36.834444 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 11 13:26:36 crc kubenswrapper[4923]: I0311 13:26:36.896119 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 11 13:26:36 crc kubenswrapper[4923]: I0311 13:26:36.935333 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 11 13:26:36 crc kubenswrapper[4923]: I0311 13:26:36.945364 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 11 13:26:36 crc kubenswrapper[4923]: I0311 13:26:36.973164 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 11 13:26:37 crc kubenswrapper[4923]: I0311 13:26:37.109329 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 11 13:26:37 crc kubenswrapper[4923]: I0311 13:26:37.111921 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 11 13:26:37 crc kubenswrapper[4923]: I0311 13:26:37.177811 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 11 13:26:37 crc kubenswrapper[4923]: I0311 13:26:37.227154 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 11 13:26:37 crc kubenswrapper[4923]: I0311 13:26:37.282437 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 11 13:26:37 crc kubenswrapper[4923]: I0311 13:26:37.288531 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 11 13:26:37 crc kubenswrapper[4923]: I0311 13:26:37.427961 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 11 13:26:37 crc kubenswrapper[4923]: I0311 13:26:37.451285 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 11 13:26:37 crc kubenswrapper[4923]: I0311 13:26:37.469900 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 11 13:26:37 crc kubenswrapper[4923]: I0311 13:26:37.528866 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 11 13:26:37 crc kubenswrapper[4923]: I0311 13:26:37.536208 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 11 13:26:37 crc kubenswrapper[4923]: I0311 13:26:37.548431 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 11 13:26:37 crc kubenswrapper[4923]: I0311 13:26:37.595054 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 11 13:26:37 crc kubenswrapper[4923]: I0311 13:26:37.695407 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 11 13:26:37 crc kubenswrapper[4923]: I0311 13:26:37.738447 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 11 13:26:37 crc kubenswrapper[4923]: I0311 13:26:37.792975 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 11 13:26:37 crc kubenswrapper[4923]: I0311 13:26:37.834001 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 11 13:26:37 crc kubenswrapper[4923]: I0311 13:26:37.897968 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 11 13:26:37 crc kubenswrapper[4923]: I0311 13:26:37.946442 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 11 13:26:38 crc kubenswrapper[4923]: I0311 13:26:38.011384 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 11 13:26:38 crc kubenswrapper[4923]: I0311 13:26:38.047526 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 11 13:26:38 crc kubenswrapper[4923]: I0311 13:26:38.171005 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 11 13:26:38 crc kubenswrapper[4923]: I0311 13:26:38.256898 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 11 13:26:38 crc kubenswrapper[4923]: I0311 13:26:38.365058 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 11 13:26:38 crc kubenswrapper[4923]: I0311 13:26:38.373644 4923 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 11 13:26:38 crc kubenswrapper[4923]: I0311 13:26:38.460731 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 11 13:26:38 crc kubenswrapper[4923]: I0311 13:26:38.500437 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 11 13:26:38 crc kubenswrapper[4923]: I0311 13:26:38.566325 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 11 13:26:38 crc kubenswrapper[4923]: I0311 13:26:38.687665 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 11 13:26:38 crc kubenswrapper[4923]: I0311 13:26:38.724188 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 11 13:26:38 crc kubenswrapper[4923]: I0311 13:26:38.804814 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 11 13:26:38 crc kubenswrapper[4923]: I0311 13:26:38.828828 4923 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 11 13:26:38 crc kubenswrapper[4923]: I0311 13:26:38.834710 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 11 13:26:38 crc kubenswrapper[4923]: I0311 13:26:38.834777 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 11 13:26:38 crc kubenswrapper[4923]: I0311 13:26:38.840068 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 11 13:26:38 crc kubenswrapper[4923]: I0311 13:26:38.859999 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=20.859981819 podStartE2EDuration="20.859981819s" podCreationTimestamp="2026-03-11 13:26:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:26:38.858831707 +0000 UTC m=+246.540733411" watchObservedRunningTime="2026-03-11 13:26:38.859981819 +0000 UTC m=+246.541883533" Mar 11 13:26:38 crc kubenswrapper[4923]: I0311 13:26:38.867323 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 11 13:26:38 crc kubenswrapper[4923]: I0311 13:26:38.942887 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 11 13:26:38 crc kubenswrapper[4923]: I0311 13:26:38.974704 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 11 13:26:38 crc kubenswrapper[4923]: I0311 13:26:38.976599 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 11 13:26:39 crc kubenswrapper[4923]: I0311 13:26:39.095304 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 11 13:26:39 crc kubenswrapper[4923]: I0311 13:26:39.129188 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 11 13:26:39 crc kubenswrapper[4923]: I0311 13:26:39.145291 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 11 13:26:39 crc kubenswrapper[4923]: I0311 13:26:39.241243 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 11 13:26:39 crc kubenswrapper[4923]: I0311 13:26:39.297114 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 11 13:26:39 crc kubenswrapper[4923]: I0311 13:26:39.306881 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 11 13:26:39 crc kubenswrapper[4923]: I0311 13:26:39.503958 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 11 13:26:39 crc kubenswrapper[4923]: I0311 13:26:39.535526 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 11 13:26:39 crc kubenswrapper[4923]: I0311 13:26:39.537867 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 11 13:26:39 crc kubenswrapper[4923]: I0311 13:26:39.580178 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 11 13:26:39 crc kubenswrapper[4923]: I0311 13:26:39.589628 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 11 13:26:39 crc kubenswrapper[4923]: I0311 13:26:39.591901 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 11 13:26:39 crc kubenswrapper[4923]: I0311 13:26:39.734747 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 11 13:26:39 crc kubenswrapper[4923]: I0311 13:26:39.753547 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 11 13:26:39 crc kubenswrapper[4923]: I0311 13:26:39.837671 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 11 13:26:39 crc kubenswrapper[4923]: I0311 13:26:39.879219 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 11 13:26:39 crc kubenswrapper[4923]: I0311 13:26:39.927243 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 11 13:26:40 crc kubenswrapper[4923]: I0311 13:26:39.999974 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 11 13:26:40 crc kubenswrapper[4923]: I0311 13:26:40.195717 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 11 13:26:40 crc kubenswrapper[4923]: I0311 13:26:40.201690 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 11 13:26:40 crc kubenswrapper[4923]: I0311 13:26:40.233280 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 11 13:26:40 crc kubenswrapper[4923]: I0311 13:26:40.238862 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 11 13:26:40 crc kubenswrapper[4923]: I0311 13:26:40.376075 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 11 13:26:40 crc kubenswrapper[4923]: I0311 13:26:40.404280 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 11 13:26:40 crc kubenswrapper[4923]: I0311 13:26:40.541648 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 11 13:26:40 crc kubenswrapper[4923]: I0311 13:26:40.577871 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 11 13:26:40 crc kubenswrapper[4923]: I0311 13:26:40.579800 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 11 13:26:40 crc kubenswrapper[4923]: I0311 13:26:40.596092 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 11 13:26:40 crc kubenswrapper[4923]: I0311 13:26:40.607773 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 11 13:26:40 crc kubenswrapper[4923]: I0311 13:26:40.871052 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 11 13:26:40 crc kubenswrapper[4923]: I0311 13:26:40.873868 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 11 13:26:40 crc kubenswrapper[4923]: I0311 13:26:40.925819 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 11 13:26:40 crc kubenswrapper[4923]: I0311 13:26:40.946135 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 11 13:26:40 crc kubenswrapper[4923]: I0311 13:26:40.974850 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 11 13:26:41 crc kubenswrapper[4923]: I0311 13:26:41.092885 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 11 13:26:41 crc kubenswrapper[4923]: I0311 13:26:41.095100 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 11 13:26:41 crc kubenswrapper[4923]: I0311 13:26:41.140141 4923 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 11 13:26:41 crc kubenswrapper[4923]: I0311 13:26:41.140444 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://8839419742b703ea5a1d3db097f4d336d55ae25bb99d9212dffa808eec14601f" gracePeriod=5 Mar 11 13:26:41 crc kubenswrapper[4923]: I0311 13:26:41.244864 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 11 13:26:41 crc kubenswrapper[4923]: I0311 13:26:41.253264 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 11 13:26:41 crc kubenswrapper[4923]: I0311 13:26:41.298221 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 11 13:26:41 crc kubenswrapper[4923]: I0311 13:26:41.309657 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 11 13:26:41 crc kubenswrapper[4923]: I0311 13:26:41.380908 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 11 13:26:41 crc kubenswrapper[4923]: I0311 13:26:41.567064 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 11 13:26:41 crc kubenswrapper[4923]: I0311 13:26:41.600802 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 11 13:26:41 crc kubenswrapper[4923]: I0311 13:26:41.601840 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 11 13:26:41 crc kubenswrapper[4923]: I0311 13:26:41.613517 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 11 13:26:41 crc kubenswrapper[4923]: I0311 13:26:41.629839 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 11 13:26:41 crc kubenswrapper[4923]: I0311 13:26:41.686413 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 11 13:26:41 crc kubenswrapper[4923]: I0311 13:26:41.714093 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 11 13:26:41 crc kubenswrapper[4923]: I0311 13:26:41.715034 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 11 13:26:41 crc kubenswrapper[4923]: I0311 13:26:41.732920 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 11 13:26:41 crc kubenswrapper[4923]: I0311 13:26:41.916887 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 11 13:26:41 crc kubenswrapper[4923]: I0311 13:26:41.932722 4923 patch_prober.go:28] interesting pod/machine-config-daemon-fsfnw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 11 13:26:41 crc kubenswrapper[4923]: I0311 13:26:41.932818 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 11 13:26:42 crc kubenswrapper[4923]: I0311 13:26:42.026865 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 11 13:26:42 crc kubenswrapper[4923]: I0311 13:26:42.037505 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 11 13:26:42 crc kubenswrapper[4923]: I0311 13:26:42.072622 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 11 13:26:42 crc kubenswrapper[4923]: I0311 13:26:42.081788 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 11 13:26:42 crc kubenswrapper[4923]: I0311 13:26:42.375874 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 11 13:26:42 crc kubenswrapper[4923]: I0311 13:26:42.388092 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 11 13:26:42 crc kubenswrapper[4923]: I0311 13:26:42.389705 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 11 13:26:42 crc kubenswrapper[4923]: I0311 13:26:42.529014 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 11 13:26:42 crc kubenswrapper[4923]: I0311 13:26:42.563671 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 11 13:26:42 crc kubenswrapper[4923]: I0311 13:26:42.564263 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 11 13:26:42 crc kubenswrapper[4923]: I0311 13:26:42.642229 4923 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 11 13:26:42 crc kubenswrapper[4923]: I0311 13:26:42.680559 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 11 13:26:42 crc kubenswrapper[4923]: I0311 13:26:42.850130 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 11 13:26:42 crc kubenswrapper[4923]: I0311 13:26:42.916749 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 11 13:26:42 crc kubenswrapper[4923]: I0311 13:26:42.916986 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 11 13:26:42 crc kubenswrapper[4923]: I0311 13:26:42.947672 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 11 13:26:42 crc kubenswrapper[4923]: I0311 13:26:42.949001 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 11 13:26:42 crc kubenswrapper[4923]: I0311 13:26:42.955557 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 11 13:26:43 crc kubenswrapper[4923]: I0311 13:26:43.197617 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 11 13:26:43 crc kubenswrapper[4923]: I0311 13:26:43.249515 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 11 13:26:43 crc kubenswrapper[4923]: I0311 13:26:43.283807 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 11 13:26:43 crc kubenswrapper[4923]: I0311 13:26:43.346900 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 11 13:26:43 crc kubenswrapper[4923]: I0311 13:26:43.361962 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 11 13:26:43 crc kubenswrapper[4923]: I0311 13:26:43.412645 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 11 13:26:43 crc kubenswrapper[4923]: I0311 13:26:43.596676 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 11 13:26:43 crc kubenswrapper[4923]: I0311 13:26:43.599731 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 11 13:26:43 crc kubenswrapper[4923]: I0311 13:26:43.766294 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 11 13:26:43 crc kubenswrapper[4923]: I0311 13:26:43.812521 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 11 13:26:43 crc kubenswrapper[4923]: I0311 13:26:43.844746 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 11 13:26:43 crc kubenswrapper[4923]: I0311 13:26:43.864511 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 11 13:26:43 crc kubenswrapper[4923]: I0311 13:26:43.930443 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 11 13:26:43 crc kubenswrapper[4923]: I0311 13:26:43.966240 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 11 13:26:43 crc kubenswrapper[4923]: I0311 13:26:43.969442 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.053375 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.105805 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.113047 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29553926-tm54x"] Mar 11 13:26:44 crc kubenswrapper[4923]: E0311 13:26:44.113515 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.113528 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 11 13:26:44 crc kubenswrapper[4923]: E0311 13:26:44.113551 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="331aaa7c-9dd6-4f2c-b3d8-00febd6448b8" containerName="installer" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.113557 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="331aaa7c-9dd6-4f2c-b3d8-00febd6448b8" containerName="installer" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.113729 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="331aaa7c-9dd6-4f2c-b3d8-00febd6448b8" containerName="installer" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.113750 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.117381 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553926-tm54x" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.120728 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.120958 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.124940 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-nxck7" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.138087 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.163977 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.172399 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj9l8\" (UniqueName: \"kubernetes.io/projected/efe004c2-3d9c-4ccb-b6c8-fbc5aa9dd615-kube-api-access-jj9l8\") pod \"auto-csr-approver-29553926-tm54x\" (UID: \"efe004c2-3d9c-4ccb-b6c8-fbc5aa9dd615\") " pod="openshift-infra/auto-csr-approver-29553926-tm54x" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.274037 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj9l8\" (UniqueName: \"kubernetes.io/projected/efe004c2-3d9c-4ccb-b6c8-fbc5aa9dd615-kube-api-access-jj9l8\") pod \"auto-csr-approver-29553926-tm54x\" (UID: \"efe004c2-3d9c-4ccb-b6c8-fbc5aa9dd615\") " pod="openshift-infra/auto-csr-approver-29553926-tm54x" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.311781 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj9l8\" (UniqueName: \"kubernetes.io/projected/efe004c2-3d9c-4ccb-b6c8-fbc5aa9dd615-kube-api-access-jj9l8\") pod \"auto-csr-approver-29553926-tm54x\" (UID: \"efe004c2-3d9c-4ccb-b6c8-fbc5aa9dd615\") " pod="openshift-infra/auto-csr-approver-29553926-tm54x" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.320173 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.385789 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.422983 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.466054 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553926-tm54x" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.468182 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.531380 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553926-tm54x"] Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.712508 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.732488 4923 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.745016 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.745162 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.865541 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 11 13:26:44 crc kubenswrapper[4923]: I0311 13:26:44.891175 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553926-tm54x"] Mar 11 13:26:45 crc kubenswrapper[4923]: I0311 13:26:45.029121 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 11 13:26:45 crc kubenswrapper[4923]: I0311 13:26:45.499512 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 11 13:26:45 crc kubenswrapper[4923]: I0311 13:26:45.633873 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 11 13:26:45 crc kubenswrapper[4923]: I0311 13:26:45.663564 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553926-tm54x" event={"ID":"efe004c2-3d9c-4ccb-b6c8-fbc5aa9dd615","Type":"ContainerStarted","Data":"ae3a6e5ef2a976d65157427f317fe321ec8607a9ac79545c4f8414e9ae36e7c1"} Mar 11 13:26:45 crc kubenswrapper[4923]: I0311 13:26:45.912314 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 11 13:26:45 crc kubenswrapper[4923]: I0311 13:26:45.939214 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 11 13:26:46 crc kubenswrapper[4923]: I0311 13:26:46.149132 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 11 13:26:46 crc kubenswrapper[4923]: I0311 13:26:46.225230 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 11 13:26:46 crc kubenswrapper[4923]: I0311 13:26:46.288313 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 11 13:26:46 crc kubenswrapper[4923]: I0311 13:26:46.440060 4923 ???:1] "http: TLS handshake error from 192.168.126.11:35860: no serving certificate available for the kubelet" Mar 11 13:26:46 crc kubenswrapper[4923]: I0311 13:26:46.626953 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 11 13:26:46 crc kubenswrapper[4923]: I0311 13:26:46.670021 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 11 13:26:46 crc kubenswrapper[4923]: I0311 13:26:46.670063 4923 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="8839419742b703ea5a1d3db097f4d336d55ae25bb99d9212dffa808eec14601f" exitCode=137 Mar 11 13:26:46 crc kubenswrapper[4923]: I0311 13:26:46.738517 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 11 13:26:46 crc kubenswrapper[4923]: I0311 13:26:46.738596 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 11 13:26:46 crc kubenswrapper[4923]: I0311 13:26:46.909323 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 11 13:26:46 crc kubenswrapper[4923]: I0311 13:26:46.909381 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 11 13:26:46 crc kubenswrapper[4923]: I0311 13:26:46.909415 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 11 13:26:46 crc kubenswrapper[4923]: I0311 13:26:46.909506 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 11 13:26:46 crc kubenswrapper[4923]: I0311 13:26:46.909531 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 11 13:26:46 crc kubenswrapper[4923]: I0311 13:26:46.909504 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:26:46 crc kubenswrapper[4923]: I0311 13:26:46.909576 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:26:46 crc kubenswrapper[4923]: I0311 13:26:46.909625 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:26:46 crc kubenswrapper[4923]: I0311 13:26:46.909736 4923 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 11 13:26:46 crc kubenswrapper[4923]: I0311 13:26:46.909735 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:26:46 crc kubenswrapper[4923]: I0311 13:26:46.909749 4923 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 11 13:26:46 crc kubenswrapper[4923]: I0311 13:26:46.909758 4923 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 11 13:26:46 crc kubenswrapper[4923]: I0311 13:26:46.916179 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:26:47 crc kubenswrapper[4923]: I0311 13:26:47.010802 4923 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 11 13:26:47 crc kubenswrapper[4923]: I0311 13:26:47.010840 4923 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 11 13:26:47 crc kubenswrapper[4923]: I0311 13:26:47.035371 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 11 13:26:47 crc kubenswrapper[4923]: I0311 13:26:47.155382 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 11 13:26:47 crc kubenswrapper[4923]: I0311 13:26:47.675569 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 11 13:26:47 crc kubenswrapper[4923]: I0311 13:26:47.675628 4923 scope.go:117] "RemoveContainer" containerID="8839419742b703ea5a1d3db097f4d336d55ae25bb99d9212dffa808eec14601f" Mar 11 13:26:47 crc kubenswrapper[4923]: I0311 13:26:47.675745 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 11 13:26:47 crc kubenswrapper[4923]: I0311 13:26:47.905903 4923 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 11 13:26:48 crc kubenswrapper[4923]: I0311 13:26:48.003890 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 11 13:26:51 crc kubenswrapper[4923]: I0311 13:26:51.533385 4923 csr.go:261] certificate signing request csr-wmhpq is approved, waiting to be issued Mar 11 13:26:51 crc kubenswrapper[4923]: I0311 13:26:51.560315 4923 csr.go:257] certificate signing request csr-wmhpq is issued Mar 11 13:26:51 crc kubenswrapper[4923]: I0311 13:26:51.703263 4923 generic.go:334] "Generic (PLEG): container finished" podID="efe004c2-3d9c-4ccb-b6c8-fbc5aa9dd615" containerID="cba92cc62a228910fdcadcc1532b801949cd6ea350cd346bd309e4fb4b3569da" exitCode=0 Mar 11 13:26:51 crc kubenswrapper[4923]: I0311 13:26:51.703307 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553926-tm54x" event={"ID":"efe004c2-3d9c-4ccb-b6c8-fbc5aa9dd615","Type":"ContainerDied","Data":"cba92cc62a228910fdcadcc1532b801949cd6ea350cd346bd309e4fb4b3569da"} Mar 11 13:26:52 crc kubenswrapper[4923]: I0311 13:26:52.562299 4923 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-18 01:41:05.211448525 +0000 UTC Mar 11 13:26:52 crc kubenswrapper[4923]: I0311 13:26:52.562337 4923 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6756h14m12.649113409s for next certificate rotation Mar 11 13:26:52 crc kubenswrapper[4923]: I0311 13:26:52.997299 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553926-tm54x" Mar 11 13:26:53 crc kubenswrapper[4923]: I0311 13:26:53.091601 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jj9l8\" (UniqueName: \"kubernetes.io/projected/efe004c2-3d9c-4ccb-b6c8-fbc5aa9dd615-kube-api-access-jj9l8\") pod \"efe004c2-3d9c-4ccb-b6c8-fbc5aa9dd615\" (UID: \"efe004c2-3d9c-4ccb-b6c8-fbc5aa9dd615\") " Mar 11 13:26:53 crc kubenswrapper[4923]: I0311 13:26:53.098659 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efe004c2-3d9c-4ccb-b6c8-fbc5aa9dd615-kube-api-access-jj9l8" (OuterVolumeSpecName: "kube-api-access-jj9l8") pod "efe004c2-3d9c-4ccb-b6c8-fbc5aa9dd615" (UID: "efe004c2-3d9c-4ccb-b6c8-fbc5aa9dd615"). InnerVolumeSpecName "kube-api-access-jj9l8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:26:53 crc kubenswrapper[4923]: I0311 13:26:53.193258 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jj9l8\" (UniqueName: \"kubernetes.io/projected/efe004c2-3d9c-4ccb-b6c8-fbc5aa9dd615-kube-api-access-jj9l8\") on node \"crc\" DevicePath \"\"" Mar 11 13:26:53 crc kubenswrapper[4923]: I0311 13:26:53.563159 4923 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-21 10:50:44.231091536 +0000 UTC Mar 11 13:26:53 crc kubenswrapper[4923]: I0311 13:26:53.563212 4923 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6837h23m50.667882889s for next certificate rotation Mar 11 13:26:53 crc kubenswrapper[4923]: I0311 13:26:53.714899 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553926-tm54x" event={"ID":"efe004c2-3d9c-4ccb-b6c8-fbc5aa9dd615","Type":"ContainerDied","Data":"ae3a6e5ef2a976d65157427f317fe321ec8607a9ac79545c4f8414e9ae36e7c1"} Mar 11 13:26:53 crc kubenswrapper[4923]: I0311 13:26:53.714936 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae3a6e5ef2a976d65157427f317fe321ec8607a9ac79545c4f8414e9ae36e7c1" Mar 11 13:26:53 crc kubenswrapper[4923]: I0311 13:26:53.714968 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553926-tm54x" Mar 11 13:27:03 crc kubenswrapper[4923]: I0311 13:27:03.772326 4923 generic.go:334] "Generic (PLEG): container finished" podID="b6887dd3-ec66-4364-8eed-f63615c35024" containerID="42bec9be6280f596ed7fc7a0fa6709d2fa9c145343570729298675bbb530a4b9" exitCode=0 Mar 11 13:27:03 crc kubenswrapper[4923]: I0311 13:27:03.772412 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" event={"ID":"b6887dd3-ec66-4364-8eed-f63615c35024","Type":"ContainerDied","Data":"42bec9be6280f596ed7fc7a0fa6709d2fa9c145343570729298675bbb530a4b9"} Mar 11 13:27:03 crc kubenswrapper[4923]: I0311 13:27:03.773578 4923 scope.go:117] "RemoveContainer" containerID="42bec9be6280f596ed7fc7a0fa6709d2fa9c145343570729298675bbb530a4b9" Mar 11 13:27:04 crc kubenswrapper[4923]: I0311 13:27:04.783174 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" event={"ID":"b6887dd3-ec66-4364-8eed-f63615c35024","Type":"ContainerStarted","Data":"e5a327ec7b02000149860befa6a02472f35beb7dec1cc5259059675ba7536f41"} Mar 11 13:27:04 crc kubenswrapper[4923]: I0311 13:27:04.784115 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" Mar 11 13:27:04 crc kubenswrapper[4923]: I0311 13:27:04.785657 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" Mar 11 13:27:11 crc kubenswrapper[4923]: I0311 13:27:11.932396 4923 patch_prober.go:28] interesting pod/machine-config-daemon-fsfnw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 11 13:27:11 crc kubenswrapper[4923]: I0311 13:27:11.933044 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 11 13:27:41 crc kubenswrapper[4923]: I0311 13:27:41.937917 4923 patch_prober.go:28] interesting pod/machine-config-daemon-fsfnw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 11 13:27:41 crc kubenswrapper[4923]: I0311 13:27:41.939419 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 11 13:27:41 crc kubenswrapper[4923]: I0311 13:27:41.940027 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" Mar 11 13:27:41 crc kubenswrapper[4923]: I0311 13:27:41.941281 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e04ff4641c7ac0b2e854199edde21a639f88c1a928f8af341f17e144a3a65e5d"} pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 11 13:27:41 crc kubenswrapper[4923]: I0311 13:27:41.941461 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" containerID="cri-o://e04ff4641c7ac0b2e854199edde21a639f88c1a928f8af341f17e144a3a65e5d" gracePeriod=600 Mar 11 13:27:43 crc kubenswrapper[4923]: I0311 13:27:43.071900 4923 generic.go:334] "Generic (PLEG): container finished" podID="356de301-ded3-4112-8406-1a28e44c2082" containerID="e04ff4641c7ac0b2e854199edde21a639f88c1a928f8af341f17e144a3a65e5d" exitCode=0 Mar 11 13:27:43 crc kubenswrapper[4923]: I0311 13:27:43.071979 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" event={"ID":"356de301-ded3-4112-8406-1a28e44c2082","Type":"ContainerDied","Data":"e04ff4641c7ac0b2e854199edde21a639f88c1a928f8af341f17e144a3a65e5d"} Mar 11 13:27:43 crc kubenswrapper[4923]: I0311 13:27:43.072602 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" event={"ID":"356de301-ded3-4112-8406-1a28e44c2082","Type":"ContainerStarted","Data":"e24d6b8a7efc75b0eaec16ef2d2e5b49a3a73ee933ce1436580a5243233d6453"} Mar 11 13:28:00 crc kubenswrapper[4923]: I0311 13:28:00.143147 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29553928-db6gx"] Mar 11 13:28:00 crc kubenswrapper[4923]: E0311 13:28:00.144236 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efe004c2-3d9c-4ccb-b6c8-fbc5aa9dd615" containerName="oc" Mar 11 13:28:00 crc kubenswrapper[4923]: I0311 13:28:00.144266 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="efe004c2-3d9c-4ccb-b6c8-fbc5aa9dd615" containerName="oc" Mar 11 13:28:00 crc kubenswrapper[4923]: I0311 13:28:00.144578 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="efe004c2-3d9c-4ccb-b6c8-fbc5aa9dd615" containerName="oc" Mar 11 13:28:00 crc kubenswrapper[4923]: I0311 13:28:00.145298 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553928-db6gx" Mar 11 13:28:00 crc kubenswrapper[4923]: I0311 13:28:00.149541 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 11 13:28:00 crc kubenswrapper[4923]: I0311 13:28:00.150064 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-nxck7" Mar 11 13:28:00 crc kubenswrapper[4923]: I0311 13:28:00.149728 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 11 13:28:00 crc kubenswrapper[4923]: I0311 13:28:00.158430 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553928-db6gx"] Mar 11 13:28:00 crc kubenswrapper[4923]: I0311 13:28:00.236054 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnkg7\" (UniqueName: \"kubernetes.io/projected/4d7c3568-b74c-4389-99e9-4bf2b62bff75-kube-api-access-gnkg7\") pod \"auto-csr-approver-29553928-db6gx\" (UID: \"4d7c3568-b74c-4389-99e9-4bf2b62bff75\") " pod="openshift-infra/auto-csr-approver-29553928-db6gx" Mar 11 13:28:00 crc kubenswrapper[4923]: I0311 13:28:00.336985 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnkg7\" (UniqueName: \"kubernetes.io/projected/4d7c3568-b74c-4389-99e9-4bf2b62bff75-kube-api-access-gnkg7\") pod \"auto-csr-approver-29553928-db6gx\" (UID: \"4d7c3568-b74c-4389-99e9-4bf2b62bff75\") " pod="openshift-infra/auto-csr-approver-29553928-db6gx" Mar 11 13:28:00 crc kubenswrapper[4923]: I0311 13:28:00.370612 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnkg7\" (UniqueName: \"kubernetes.io/projected/4d7c3568-b74c-4389-99e9-4bf2b62bff75-kube-api-access-gnkg7\") pod \"auto-csr-approver-29553928-db6gx\" (UID: \"4d7c3568-b74c-4389-99e9-4bf2b62bff75\") " pod="openshift-infra/auto-csr-approver-29553928-db6gx" Mar 11 13:28:00 crc kubenswrapper[4923]: I0311 13:28:00.523421 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553928-db6gx" Mar 11 13:28:00 crc kubenswrapper[4923]: I0311 13:28:00.992812 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553928-db6gx"] Mar 11 13:28:01 crc kubenswrapper[4923]: I0311 13:28:01.185466 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553928-db6gx" event={"ID":"4d7c3568-b74c-4389-99e9-4bf2b62bff75","Type":"ContainerStarted","Data":"022e8c00e070bfaccaf459e36ef5ba27a97ae273c988f0f514484ba5cfba6515"} Mar 11 13:28:03 crc kubenswrapper[4923]: I0311 13:28:03.267873 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553928-db6gx" event={"ID":"4d7c3568-b74c-4389-99e9-4bf2b62bff75","Type":"ContainerStarted","Data":"46042e4f6d77fb1afb0946f9286c6e087113489c7474adf289f0c1f5aeba6f4f"} Mar 11 13:28:04 crc kubenswrapper[4923]: I0311 13:28:04.280415 4923 generic.go:334] "Generic (PLEG): container finished" podID="4d7c3568-b74c-4389-99e9-4bf2b62bff75" containerID="46042e4f6d77fb1afb0946f9286c6e087113489c7474adf289f0c1f5aeba6f4f" exitCode=0 Mar 11 13:28:04 crc kubenswrapper[4923]: I0311 13:28:04.280480 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553928-db6gx" event={"ID":"4d7c3568-b74c-4389-99e9-4bf2b62bff75","Type":"ContainerDied","Data":"46042e4f6d77fb1afb0946f9286c6e087113489c7474adf289f0c1f5aeba6f4f"} Mar 11 13:28:05 crc kubenswrapper[4923]: I0311 13:28:05.489448 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553928-db6gx" Mar 11 13:28:05 crc kubenswrapper[4923]: I0311 13:28:05.678204 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnkg7\" (UniqueName: \"kubernetes.io/projected/4d7c3568-b74c-4389-99e9-4bf2b62bff75-kube-api-access-gnkg7\") pod \"4d7c3568-b74c-4389-99e9-4bf2b62bff75\" (UID: \"4d7c3568-b74c-4389-99e9-4bf2b62bff75\") " Mar 11 13:28:05 crc kubenswrapper[4923]: I0311 13:28:05.686547 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d7c3568-b74c-4389-99e9-4bf2b62bff75-kube-api-access-gnkg7" (OuterVolumeSpecName: "kube-api-access-gnkg7") pod "4d7c3568-b74c-4389-99e9-4bf2b62bff75" (UID: "4d7c3568-b74c-4389-99e9-4bf2b62bff75"). InnerVolumeSpecName "kube-api-access-gnkg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:28:05 crc kubenswrapper[4923]: I0311 13:28:05.779955 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnkg7\" (UniqueName: \"kubernetes.io/projected/4d7c3568-b74c-4389-99e9-4bf2b62bff75-kube-api-access-gnkg7\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:06 crc kubenswrapper[4923]: I0311 13:28:06.293686 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553928-db6gx" event={"ID":"4d7c3568-b74c-4389-99e9-4bf2b62bff75","Type":"ContainerDied","Data":"022e8c00e070bfaccaf459e36ef5ba27a97ae273c988f0f514484ba5cfba6515"} Mar 11 13:28:06 crc kubenswrapper[4923]: I0311 13:28:06.294121 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="022e8c00e070bfaccaf459e36ef5ba27a97ae273c988f0f514484ba5cfba6515" Mar 11 13:28:06 crc kubenswrapper[4923]: I0311 13:28:06.293752 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553928-db6gx" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.251279 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-5zbtq"] Mar 11 13:28:13 crc kubenswrapper[4923]: E0311 13:28:13.252032 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d7c3568-b74c-4389-99e9-4bf2b62bff75" containerName="oc" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.252044 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d7c3568-b74c-4389-99e9-4bf2b62bff75" containerName="oc" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.252140 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d7c3568-b74c-4389-99e9-4bf2b62bff75" containerName="oc" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.252529 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.271583 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-5zbtq"] Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.380890 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swgnb\" (UniqueName: \"kubernetes.io/projected/51826efa-49b4-44ba-a988-5b59ca72165b-kube-api-access-swgnb\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.380934 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/51826efa-49b4-44ba-a988-5b59ca72165b-registry-certificates\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.380966 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/51826efa-49b4-44ba-a988-5b59ca72165b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.380992 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.381018 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/51826efa-49b4-44ba-a988-5b59ca72165b-trusted-ca\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.381058 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/51826efa-49b4-44ba-a988-5b59ca72165b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.381083 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/51826efa-49b4-44ba-a988-5b59ca72165b-bound-sa-token\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.381111 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/51826efa-49b4-44ba-a988-5b59ca72165b-registry-tls\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.402169 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.482324 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swgnb\" (UniqueName: \"kubernetes.io/projected/51826efa-49b4-44ba-a988-5b59ca72165b-kube-api-access-swgnb\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.482393 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/51826efa-49b4-44ba-a988-5b59ca72165b-registry-certificates\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.482426 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/51826efa-49b4-44ba-a988-5b59ca72165b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.482447 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/51826efa-49b4-44ba-a988-5b59ca72165b-trusted-ca\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.482490 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/51826efa-49b4-44ba-a988-5b59ca72165b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.482510 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/51826efa-49b4-44ba-a988-5b59ca72165b-bound-sa-token\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.482533 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/51826efa-49b4-44ba-a988-5b59ca72165b-registry-tls\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.483203 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/51826efa-49b4-44ba-a988-5b59ca72165b-ca-trust-extracted\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.483678 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/51826efa-49b4-44ba-a988-5b59ca72165b-trusted-ca\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.483791 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/51826efa-49b4-44ba-a988-5b59ca72165b-registry-certificates\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.491279 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/51826efa-49b4-44ba-a988-5b59ca72165b-installation-pull-secrets\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.491312 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/51826efa-49b4-44ba-a988-5b59ca72165b-registry-tls\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.502844 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swgnb\" (UniqueName: \"kubernetes.io/projected/51826efa-49b4-44ba-a988-5b59ca72165b-kube-api-access-swgnb\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.504194 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/51826efa-49b4-44ba-a988-5b59ca72165b-bound-sa-token\") pod \"image-registry-66df7c8f76-5zbtq\" (UID: \"51826efa-49b4-44ba-a988-5b59ca72165b\") " pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.567839 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:13 crc kubenswrapper[4923]: I0311 13:28:13.768779 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-5zbtq"] Mar 11 13:28:14 crc kubenswrapper[4923]: I0311 13:28:14.340603 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" event={"ID":"51826efa-49b4-44ba-a988-5b59ca72165b","Type":"ContainerStarted","Data":"d8c52ded177c8532c24c1d8cb783258feb3ef10859cd1b8fb37b9438f8bd0b6a"} Mar 11 13:28:14 crc kubenswrapper[4923]: I0311 13:28:14.341229 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" event={"ID":"51826efa-49b4-44ba-a988-5b59ca72165b","Type":"ContainerStarted","Data":"d233416c6a425e58797017783c089c0f793fab684030baf0c17088ff8919f62d"} Mar 11 13:28:14 crc kubenswrapper[4923]: I0311 13:28:14.341248 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:14 crc kubenswrapper[4923]: I0311 13:28:14.374253 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" podStartSLOduration=1.374225254 podStartE2EDuration="1.374225254s" podCreationTimestamp="2026-03-11 13:28:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:28:14.368195469 +0000 UTC m=+342.050097193" watchObservedRunningTime="2026-03-11 13:28:14.374225254 +0000 UTC m=+342.056127008" Mar 11 13:28:14 crc kubenswrapper[4923]: I0311 13:28:14.968222 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8f4cb"] Mar 11 13:28:14 crc kubenswrapper[4923]: I0311 13:28:14.968691 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8f4cb" podUID="52f59385-23ad-4163-a766-b9d21e57ad6c" containerName="registry-server" containerID="cri-o://784602a924d09abfb2e1a681cacf54d66ca459bb02d50b9e2b24ba71177c485a" gracePeriod=30 Mar 11 13:28:14 crc kubenswrapper[4923]: I0311 13:28:14.978873 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cqdq8"] Mar 11 13:28:14 crc kubenswrapper[4923]: I0311 13:28:14.979150 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cqdq8" podUID="7ce0bb4a-0841-4449-bc02-6c219731272e" containerName="registry-server" containerID="cri-o://3e492da31d96171c02f8879c5e7423f7a1c454b1fc4d109a1c5db5ab6c4a2c0c" gracePeriod=30 Mar 11 13:28:14 crc kubenswrapper[4923]: I0311 13:28:14.989180 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vznqg"] Mar 11 13:28:14 crc kubenswrapper[4923]: I0311 13:28:14.989447 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" podUID="b6887dd3-ec66-4364-8eed-f63615c35024" containerName="marketplace-operator" containerID="cri-o://e5a327ec7b02000149860befa6a02472f35beb7dec1cc5259059675ba7536f41" gracePeriod=30 Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.001808 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jcncp"] Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.002187 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jcncp" podUID="ac5c84a5-f129-440d-8423-f8d86b00e3b9" containerName="registry-server" containerID="cri-o://97442e068883c28e9c46d4062a84e4429f6597de785028bde4b003dfc13d63d6" gracePeriod=30 Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.015908 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z7n2v"] Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.017029 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z7n2v" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.019220 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rnjrh"] Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.019496 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rnjrh" podUID="094d43b0-8aa7-4acc-9d97-9c00874ee1a8" containerName="registry-server" containerID="cri-o://fe4253ba3094d50d1f38cf57ba1844808af2c64277e0140ca9e55de3afdd26a8" gracePeriod=30 Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.038258 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z7n2v"] Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.212034 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/85ba46fa-fad0-421a-8b0a-e83a8e417079-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z7n2v\" (UID: \"85ba46fa-fad0-421a-8b0a-e83a8e417079\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7n2v" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.212079 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx4qx\" (UniqueName: \"kubernetes.io/projected/85ba46fa-fad0-421a-8b0a-e83a8e417079-kube-api-access-qx4qx\") pod \"marketplace-operator-79b997595-z7n2v\" (UID: \"85ba46fa-fad0-421a-8b0a-e83a8e417079\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7n2v" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.212124 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/85ba46fa-fad0-421a-8b0a-e83a8e417079-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z7n2v\" (UID: \"85ba46fa-fad0-421a-8b0a-e83a8e417079\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7n2v" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.313670 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/85ba46fa-fad0-421a-8b0a-e83a8e417079-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z7n2v\" (UID: \"85ba46fa-fad0-421a-8b0a-e83a8e417079\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7n2v" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.313719 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx4qx\" (UniqueName: \"kubernetes.io/projected/85ba46fa-fad0-421a-8b0a-e83a8e417079-kube-api-access-qx4qx\") pod \"marketplace-operator-79b997595-z7n2v\" (UID: \"85ba46fa-fad0-421a-8b0a-e83a8e417079\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7n2v" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.313770 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/85ba46fa-fad0-421a-8b0a-e83a8e417079-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z7n2v\" (UID: \"85ba46fa-fad0-421a-8b0a-e83a8e417079\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7n2v" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.315080 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/85ba46fa-fad0-421a-8b0a-e83a8e417079-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-z7n2v\" (UID: \"85ba46fa-fad0-421a-8b0a-e83a8e417079\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7n2v" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.319333 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/85ba46fa-fad0-421a-8b0a-e83a8e417079-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-z7n2v\" (UID: \"85ba46fa-fad0-421a-8b0a-e83a8e417079\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7n2v" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.345124 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx4qx\" (UniqueName: \"kubernetes.io/projected/85ba46fa-fad0-421a-8b0a-e83a8e417079-kube-api-access-qx4qx\") pod \"marketplace-operator-79b997595-z7n2v\" (UID: \"85ba46fa-fad0-421a-8b0a-e83a8e417079\") " pod="openshift-marketplace/marketplace-operator-79b997595-z7n2v" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.351583 4923 generic.go:334] "Generic (PLEG): container finished" podID="094d43b0-8aa7-4acc-9d97-9c00874ee1a8" containerID="fe4253ba3094d50d1f38cf57ba1844808af2c64277e0140ca9e55de3afdd26a8" exitCode=0 Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.351653 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnjrh" event={"ID":"094d43b0-8aa7-4acc-9d97-9c00874ee1a8","Type":"ContainerDied","Data":"fe4253ba3094d50d1f38cf57ba1844808af2c64277e0140ca9e55de3afdd26a8"} Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.353118 4923 generic.go:334] "Generic (PLEG): container finished" podID="7ce0bb4a-0841-4449-bc02-6c219731272e" containerID="3e492da31d96171c02f8879c5e7423f7a1c454b1fc4d109a1c5db5ab6c4a2c0c" exitCode=0 Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.353195 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqdq8" event={"ID":"7ce0bb4a-0841-4449-bc02-6c219731272e","Type":"ContainerDied","Data":"3e492da31d96171c02f8879c5e7423f7a1c454b1fc4d109a1c5db5ab6c4a2c0c"} Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.356798 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" event={"ID":"b6887dd3-ec66-4364-8eed-f63615c35024","Type":"ContainerDied","Data":"e5a327ec7b02000149860befa6a02472f35beb7dec1cc5259059675ba7536f41"} Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.356846 4923 scope.go:117] "RemoveContainer" containerID="42bec9be6280f596ed7fc7a0fa6709d2fa9c145343570729298675bbb530a4b9" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.354841 4923 generic.go:334] "Generic (PLEG): container finished" podID="b6887dd3-ec66-4364-8eed-f63615c35024" containerID="e5a327ec7b02000149860befa6a02472f35beb7dec1cc5259059675ba7536f41" exitCode=0 Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.364104 4923 generic.go:334] "Generic (PLEG): container finished" podID="52f59385-23ad-4163-a766-b9d21e57ad6c" containerID="784602a924d09abfb2e1a681cacf54d66ca459bb02d50b9e2b24ba71177c485a" exitCode=0 Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.364186 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8f4cb" event={"ID":"52f59385-23ad-4163-a766-b9d21e57ad6c","Type":"ContainerDied","Data":"784602a924d09abfb2e1a681cacf54d66ca459bb02d50b9e2b24ba71177c485a"} Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.364222 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8f4cb" event={"ID":"52f59385-23ad-4163-a766-b9d21e57ad6c","Type":"ContainerDied","Data":"dcf2b128619685189cc795ffc1f620befcc350a4ec4d6de3e7a6de4148822861"} Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.364236 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcf2b128619685189cc795ffc1f620befcc350a4ec4d6de3e7a6de4148822861" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.371328 4923 generic.go:334] "Generic (PLEG): container finished" podID="ac5c84a5-f129-440d-8423-f8d86b00e3b9" containerID="97442e068883c28e9c46d4062a84e4429f6597de785028bde4b003dfc13d63d6" exitCode=0 Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.371670 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jcncp" event={"ID":"ac5c84a5-f129-440d-8423-f8d86b00e3b9","Type":"ContainerDied","Data":"97442e068883c28e9c46d4062a84e4429f6597de785028bde4b003dfc13d63d6"} Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.474132 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8f4cb" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.509183 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cqdq8" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.518632 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rnjrh" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.524883 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jcncp" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.526948 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.622811 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac5c84a5-f129-440d-8423-f8d86b00e3b9-catalog-content\") pod \"ac5c84a5-f129-440d-8423-f8d86b00e3b9\" (UID: \"ac5c84a5-f129-440d-8423-f8d86b00e3b9\") " Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.622860 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ce0bb4a-0841-4449-bc02-6c219731272e-utilities\") pod \"7ce0bb4a-0841-4449-bc02-6c219731272e\" (UID: \"7ce0bb4a-0841-4449-bc02-6c219731272e\") " Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.622887 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52f59385-23ad-4163-a766-b9d21e57ad6c-catalog-content\") pod \"52f59385-23ad-4163-a766-b9d21e57ad6c\" (UID: \"52f59385-23ad-4163-a766-b9d21e57ad6c\") " Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.622926 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6887dd3-ec66-4364-8eed-f63615c35024-marketplace-trusted-ca\") pod \"b6887dd3-ec66-4364-8eed-f63615c35024\" (UID: \"b6887dd3-ec66-4364-8eed-f63615c35024\") " Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.622948 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gfx4\" (UniqueName: \"kubernetes.io/projected/b6887dd3-ec66-4364-8eed-f63615c35024-kube-api-access-9gfx4\") pod \"b6887dd3-ec66-4364-8eed-f63615c35024\" (UID: \"b6887dd3-ec66-4364-8eed-f63615c35024\") " Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.622983 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x55th\" (UniqueName: \"kubernetes.io/projected/094d43b0-8aa7-4acc-9d97-9c00874ee1a8-kube-api-access-x55th\") pod \"094d43b0-8aa7-4acc-9d97-9c00874ee1a8\" (UID: \"094d43b0-8aa7-4acc-9d97-9c00874ee1a8\") " Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.622997 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d48t2\" (UniqueName: \"kubernetes.io/projected/52f59385-23ad-4163-a766-b9d21e57ad6c-kube-api-access-d48t2\") pod \"52f59385-23ad-4163-a766-b9d21e57ad6c\" (UID: \"52f59385-23ad-4163-a766-b9d21e57ad6c\") " Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.623017 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac5c84a5-f129-440d-8423-f8d86b00e3b9-utilities\") pod \"ac5c84a5-f129-440d-8423-f8d86b00e3b9\" (UID: \"ac5c84a5-f129-440d-8423-f8d86b00e3b9\") " Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.623031 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vq82b\" (UniqueName: \"kubernetes.io/projected/ac5c84a5-f129-440d-8423-f8d86b00e3b9-kube-api-access-vq82b\") pod \"ac5c84a5-f129-440d-8423-f8d86b00e3b9\" (UID: \"ac5c84a5-f129-440d-8423-f8d86b00e3b9\") " Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.623048 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/094d43b0-8aa7-4acc-9d97-9c00874ee1a8-catalog-content\") pod \"094d43b0-8aa7-4acc-9d97-9c00874ee1a8\" (UID: \"094d43b0-8aa7-4acc-9d97-9c00874ee1a8\") " Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.623071 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52f59385-23ad-4163-a766-b9d21e57ad6c-utilities\") pod \"52f59385-23ad-4163-a766-b9d21e57ad6c\" (UID: \"52f59385-23ad-4163-a766-b9d21e57ad6c\") " Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.623087 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ce0bb4a-0841-4449-bc02-6c219731272e-catalog-content\") pod \"7ce0bb4a-0841-4449-bc02-6c219731272e\" (UID: \"7ce0bb4a-0841-4449-bc02-6c219731272e\") " Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.623106 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mzqq\" (UniqueName: \"kubernetes.io/projected/7ce0bb4a-0841-4449-bc02-6c219731272e-kube-api-access-8mzqq\") pod \"7ce0bb4a-0841-4449-bc02-6c219731272e\" (UID: \"7ce0bb4a-0841-4449-bc02-6c219731272e\") " Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.623130 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/094d43b0-8aa7-4acc-9d97-9c00874ee1a8-utilities\") pod \"094d43b0-8aa7-4acc-9d97-9c00874ee1a8\" (UID: \"094d43b0-8aa7-4acc-9d97-9c00874ee1a8\") " Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.623150 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6887dd3-ec66-4364-8eed-f63615c35024-marketplace-operator-metrics\") pod \"b6887dd3-ec66-4364-8eed-f63615c35024\" (UID: \"b6887dd3-ec66-4364-8eed-f63615c35024\") " Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.625066 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52f59385-23ad-4163-a766-b9d21e57ad6c-utilities" (OuterVolumeSpecName: "utilities") pod "52f59385-23ad-4163-a766-b9d21e57ad6c" (UID: "52f59385-23ad-4163-a766-b9d21e57ad6c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.626444 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac5c84a5-f129-440d-8423-f8d86b00e3b9-utilities" (OuterVolumeSpecName: "utilities") pod "ac5c84a5-f129-440d-8423-f8d86b00e3b9" (UID: "ac5c84a5-f129-440d-8423-f8d86b00e3b9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.626477 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6887dd3-ec66-4364-8eed-f63615c35024-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6887dd3-ec66-4364-8eed-f63615c35024" (UID: "b6887dd3-ec66-4364-8eed-f63615c35024"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.627608 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/094d43b0-8aa7-4acc-9d97-9c00874ee1a8-utilities" (OuterVolumeSpecName: "utilities") pod "094d43b0-8aa7-4acc-9d97-9c00874ee1a8" (UID: "094d43b0-8aa7-4acc-9d97-9c00874ee1a8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.627621 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ce0bb4a-0841-4449-bc02-6c219731272e-utilities" (OuterVolumeSpecName: "utilities") pod "7ce0bb4a-0841-4449-bc02-6c219731272e" (UID: "7ce0bb4a-0841-4449-bc02-6c219731272e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.628257 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac5c84a5-f129-440d-8423-f8d86b00e3b9-kube-api-access-vq82b" (OuterVolumeSpecName: "kube-api-access-vq82b") pod "ac5c84a5-f129-440d-8423-f8d86b00e3b9" (UID: "ac5c84a5-f129-440d-8423-f8d86b00e3b9"). InnerVolumeSpecName "kube-api-access-vq82b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.628465 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/094d43b0-8aa7-4acc-9d97-9c00874ee1a8-kube-api-access-x55th" (OuterVolumeSpecName: "kube-api-access-x55th") pod "094d43b0-8aa7-4acc-9d97-9c00874ee1a8" (UID: "094d43b0-8aa7-4acc-9d97-9c00874ee1a8"). InnerVolumeSpecName "kube-api-access-x55th". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.628527 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6887dd3-ec66-4364-8eed-f63615c35024-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6887dd3-ec66-4364-8eed-f63615c35024" (UID: "b6887dd3-ec66-4364-8eed-f63615c35024"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.629264 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52f59385-23ad-4163-a766-b9d21e57ad6c-kube-api-access-d48t2" (OuterVolumeSpecName: "kube-api-access-d48t2") pod "52f59385-23ad-4163-a766-b9d21e57ad6c" (UID: "52f59385-23ad-4163-a766-b9d21e57ad6c"). InnerVolumeSpecName "kube-api-access-d48t2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.631506 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ce0bb4a-0841-4449-bc02-6c219731272e-kube-api-access-8mzqq" (OuterVolumeSpecName: "kube-api-access-8mzqq") pod "7ce0bb4a-0841-4449-bc02-6c219731272e" (UID: "7ce0bb4a-0841-4449-bc02-6c219731272e"). InnerVolumeSpecName "kube-api-access-8mzqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.636174 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-z7n2v" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.640803 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6887dd3-ec66-4364-8eed-f63615c35024-kube-api-access-9gfx4" (OuterVolumeSpecName: "kube-api-access-9gfx4") pod "b6887dd3-ec66-4364-8eed-f63615c35024" (UID: "b6887dd3-ec66-4364-8eed-f63615c35024"). InnerVolumeSpecName "kube-api-access-9gfx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.672680 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac5c84a5-f129-440d-8423-f8d86b00e3b9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac5c84a5-f129-440d-8423-f8d86b00e3b9" (UID: "ac5c84a5-f129-440d-8423-f8d86b00e3b9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.696684 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/52f59385-23ad-4163-a766-b9d21e57ad6c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "52f59385-23ad-4163-a766-b9d21e57ad6c" (UID: "52f59385-23ad-4163-a766-b9d21e57ad6c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.701722 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ce0bb4a-0841-4449-bc02-6c219731272e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7ce0bb4a-0841-4449-bc02-6c219731272e" (UID: "7ce0bb4a-0841-4449-bc02-6c219731272e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.724176 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x55th\" (UniqueName: \"kubernetes.io/projected/094d43b0-8aa7-4acc-9d97-9c00874ee1a8-kube-api-access-x55th\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.724211 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d48t2\" (UniqueName: \"kubernetes.io/projected/52f59385-23ad-4163-a766-b9d21e57ad6c-kube-api-access-d48t2\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.724224 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac5c84a5-f129-440d-8423-f8d86b00e3b9-utilities\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.724236 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vq82b\" (UniqueName: \"kubernetes.io/projected/ac5c84a5-f129-440d-8423-f8d86b00e3b9-kube-api-access-vq82b\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.724247 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/52f59385-23ad-4163-a766-b9d21e57ad6c-utilities\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.724257 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7ce0bb4a-0841-4449-bc02-6c219731272e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.724268 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mzqq\" (UniqueName: \"kubernetes.io/projected/7ce0bb4a-0841-4449-bc02-6c219731272e-kube-api-access-8mzqq\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.724279 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/094d43b0-8aa7-4acc-9d97-9c00874ee1a8-utilities\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.724290 4923 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6887dd3-ec66-4364-8eed-f63615c35024-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.724302 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac5c84a5-f129-440d-8423-f8d86b00e3b9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.724312 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7ce0bb4a-0841-4449-bc02-6c219731272e-utilities\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.724321 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/52f59385-23ad-4163-a766-b9d21e57ad6c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.724329 4923 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6887dd3-ec66-4364-8eed-f63615c35024-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.724351 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gfx4\" (UniqueName: \"kubernetes.io/projected/b6887dd3-ec66-4364-8eed-f63615c35024-kube-api-access-9gfx4\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.780660 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/094d43b0-8aa7-4acc-9d97-9c00874ee1a8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "094d43b0-8aa7-4acc-9d97-9c00874ee1a8" (UID: "094d43b0-8aa7-4acc-9d97-9c00874ee1a8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.825417 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/094d43b0-8aa7-4acc-9d97-9c00874ee1a8-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:15 crc kubenswrapper[4923]: I0311 13:28:15.839024 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-z7n2v"] Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.378275 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" event={"ID":"b6887dd3-ec66-4364-8eed-f63615c35024","Type":"ContainerDied","Data":"1fb0655229a7efa4dbd0890f27dedc82b94c6421088752b7e02c607fed31875b"} Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.378645 4923 scope.go:117] "RemoveContainer" containerID="e5a327ec7b02000149860befa6a02472f35beb7dec1cc5259059675ba7536f41" Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.378540 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vznqg" Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.381888 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jcncp" event={"ID":"ac5c84a5-f129-440d-8423-f8d86b00e3b9","Type":"ContainerDied","Data":"f66f05bc927a40ada2f03065ebbcf30bc630963cfe8ca94d4e96130f781fd07d"} Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.382010 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jcncp" Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.389711 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cqdq8" event={"ID":"7ce0bb4a-0841-4449-bc02-6c219731272e","Type":"ContainerDied","Data":"0f0311972eaa16edc13729f5c146404ff3a9210c4764b51789aa37602d767169"} Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.389796 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cqdq8" Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.398031 4923 scope.go:117] "RemoveContainer" containerID="97442e068883c28e9c46d4062a84e4429f6597de785028bde4b003dfc13d63d6" Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.399367 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rnjrh" event={"ID":"094d43b0-8aa7-4acc-9d97-9c00874ee1a8","Type":"ContainerDied","Data":"58bf1236c788c6461942b987737349e09a6c530f01f072baa7e67713d554af3f"} Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.399456 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rnjrh" Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.401146 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8f4cb" Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.407610 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z7n2v" event={"ID":"85ba46fa-fad0-421a-8b0a-e83a8e417079","Type":"ContainerStarted","Data":"51984700fbfc4eb6c120bebbcd97f778197742abfd7d7ba8c0bcea82e3a04c80"} Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.407675 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-z7n2v" Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.407696 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-z7n2v" event={"ID":"85ba46fa-fad0-421a-8b0a-e83a8e417079","Type":"ContainerStarted","Data":"473372479d67b7b64e49130a1ab03d9eb7c84ec4f43b96456c24590ce5ccb0d0"} Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.419429 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-z7n2v" Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.433418 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-z7n2v" podStartSLOduration=2.433389288 podStartE2EDuration="2.433389288s" podCreationTimestamp="2026-03-11 13:28:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:28:16.433212934 +0000 UTC m=+344.115114668" watchObservedRunningTime="2026-03-11 13:28:16.433389288 +0000 UTC m=+344.115291032" Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.442483 4923 scope.go:117] "RemoveContainer" containerID="32c9a726e5c0ba79943abfa9da284d8ef2f3e0beebc78ef5904e35d21a6c483b" Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.493067 4923 scope.go:117] "RemoveContainer" containerID="57a50dd1eeb8b3ff16a98538327c923e7ed49047fea234d6cdbe1d8c02bd4903" Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.504099 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vznqg"] Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.513741 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vznqg"] Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.514103 4923 scope.go:117] "RemoveContainer" containerID="3e492da31d96171c02f8879c5e7423f7a1c454b1fc4d109a1c5db5ab6c4a2c0c" Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.518438 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cqdq8"] Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.523110 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cqdq8"] Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.529272 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8f4cb"] Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.533013 4923 scope.go:117] "RemoveContainer" containerID="a5489594d4199b5105fab493c51e3da64bd6c7c739f44779631fd6d76a22972a" Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.534585 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8f4cb"] Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.539465 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rnjrh"] Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.543173 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rnjrh"] Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.547805 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jcncp"] Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.550717 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jcncp"] Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.556749 4923 scope.go:117] "RemoveContainer" containerID="1dde84cc50f06c1d02f0b31364a095423f98d2acad51e1a2cb5d8e255513a212" Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.575293 4923 scope.go:117] "RemoveContainer" containerID="fe4253ba3094d50d1f38cf57ba1844808af2c64277e0140ca9e55de3afdd26a8" Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.592088 4923 scope.go:117] "RemoveContainer" containerID="786a32861029cbd731746540a5ae3701ac6ca72dc83eff5f0beca46f6037f092" Mar 11 13:28:16 crc kubenswrapper[4923]: I0311 13:28:16.608607 4923 scope.go:117] "RemoveContainer" containerID="d78a8ffda1ef27c83eadbf30b6f491cfb8d2682a01577dfabdd6a0165e9db445" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.031804 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="094d43b0-8aa7-4acc-9d97-9c00874ee1a8" path="/var/lib/kubelet/pods/094d43b0-8aa7-4acc-9d97-9c00874ee1a8/volumes" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.032790 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52f59385-23ad-4163-a766-b9d21e57ad6c" path="/var/lib/kubelet/pods/52f59385-23ad-4163-a766-b9d21e57ad6c/volumes" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.033790 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ce0bb4a-0841-4449-bc02-6c219731272e" path="/var/lib/kubelet/pods/7ce0bb4a-0841-4449-bc02-6c219731272e/volumes" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.035116 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac5c84a5-f129-440d-8423-f8d86b00e3b9" path="/var/lib/kubelet/pods/ac5c84a5-f129-440d-8423-f8d86b00e3b9/volumes" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.035883 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6887dd3-ec66-4364-8eed-f63615c35024" path="/var/lib/kubelet/pods/b6887dd3-ec66-4364-8eed-f63615c35024/volumes" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.581245 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-48q6k"] Mar 11 13:28:17 crc kubenswrapper[4923]: E0311 13:28:17.582251 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52f59385-23ad-4163-a766-b9d21e57ad6c" containerName="extract-content" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.582335 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="52f59385-23ad-4163-a766-b9d21e57ad6c" containerName="extract-content" Mar 11 13:28:17 crc kubenswrapper[4923]: E0311 13:28:17.582421 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="094d43b0-8aa7-4acc-9d97-9c00874ee1a8" containerName="extract-utilities" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.582509 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="094d43b0-8aa7-4acc-9d97-9c00874ee1a8" containerName="extract-utilities" Mar 11 13:28:17 crc kubenswrapper[4923]: E0311 13:28:17.582577 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac5c84a5-f129-440d-8423-f8d86b00e3b9" containerName="extract-utilities" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.582636 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac5c84a5-f129-440d-8423-f8d86b00e3b9" containerName="extract-utilities" Mar 11 13:28:17 crc kubenswrapper[4923]: E0311 13:28:17.582693 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6887dd3-ec66-4364-8eed-f63615c35024" containerName="marketplace-operator" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.582757 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6887dd3-ec66-4364-8eed-f63615c35024" containerName="marketplace-operator" Mar 11 13:28:17 crc kubenswrapper[4923]: E0311 13:28:17.582831 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce0bb4a-0841-4449-bc02-6c219731272e" containerName="extract-content" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.582905 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce0bb4a-0841-4449-bc02-6c219731272e" containerName="extract-content" Mar 11 13:28:17 crc kubenswrapper[4923]: E0311 13:28:17.582989 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="094d43b0-8aa7-4acc-9d97-9c00874ee1a8" containerName="extract-content" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.583047 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="094d43b0-8aa7-4acc-9d97-9c00874ee1a8" containerName="extract-content" Mar 11 13:28:17 crc kubenswrapper[4923]: E0311 13:28:17.583141 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce0bb4a-0841-4449-bc02-6c219731272e" containerName="registry-server" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.583210 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce0bb4a-0841-4449-bc02-6c219731272e" containerName="registry-server" Mar 11 13:28:17 crc kubenswrapper[4923]: E0311 13:28:17.583334 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6887dd3-ec66-4364-8eed-f63615c35024" containerName="marketplace-operator" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.583408 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6887dd3-ec66-4364-8eed-f63615c35024" containerName="marketplace-operator" Mar 11 13:28:17 crc kubenswrapper[4923]: E0311 13:28:17.583464 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac5c84a5-f129-440d-8423-f8d86b00e3b9" containerName="registry-server" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.583524 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac5c84a5-f129-440d-8423-f8d86b00e3b9" containerName="registry-server" Mar 11 13:28:17 crc kubenswrapper[4923]: E0311 13:28:17.583583 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ce0bb4a-0841-4449-bc02-6c219731272e" containerName="extract-utilities" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.583633 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ce0bb4a-0841-4449-bc02-6c219731272e" containerName="extract-utilities" Mar 11 13:28:17 crc kubenswrapper[4923]: E0311 13:28:17.583688 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac5c84a5-f129-440d-8423-f8d86b00e3b9" containerName="extract-content" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.583745 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac5c84a5-f129-440d-8423-f8d86b00e3b9" containerName="extract-content" Mar 11 13:28:17 crc kubenswrapper[4923]: E0311 13:28:17.583801 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52f59385-23ad-4163-a766-b9d21e57ad6c" containerName="extract-utilities" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.583860 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="52f59385-23ad-4163-a766-b9d21e57ad6c" containerName="extract-utilities" Mar 11 13:28:17 crc kubenswrapper[4923]: E0311 13:28:17.583913 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="094d43b0-8aa7-4acc-9d97-9c00874ee1a8" containerName="registry-server" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.583964 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="094d43b0-8aa7-4acc-9d97-9c00874ee1a8" containerName="registry-server" Mar 11 13:28:17 crc kubenswrapper[4923]: E0311 13:28:17.584021 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="52f59385-23ad-4163-a766-b9d21e57ad6c" containerName="registry-server" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.584073 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="52f59385-23ad-4163-a766-b9d21e57ad6c" containerName="registry-server" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.584205 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6887dd3-ec66-4364-8eed-f63615c35024" containerName="marketplace-operator" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.584352 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac5c84a5-f129-440d-8423-f8d86b00e3b9" containerName="registry-server" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.584416 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ce0bb4a-0841-4449-bc02-6c219731272e" containerName="registry-server" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.584481 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="094d43b0-8aa7-4acc-9d97-9c00874ee1a8" containerName="registry-server" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.584535 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="52f59385-23ad-4163-a766-b9d21e57ad6c" containerName="registry-server" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.584752 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6887dd3-ec66-4364-8eed-f63615c35024" containerName="marketplace-operator" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.585387 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48q6k" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.587800 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.597507 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-48q6k"] Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.753102 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ng4v\" (UniqueName: \"kubernetes.io/projected/99a57330-6112-4e39-aec6-3f62ff4a56ea-kube-api-access-6ng4v\") pod \"redhat-operators-48q6k\" (UID: \"99a57330-6112-4e39-aec6-3f62ff4a56ea\") " pod="openshift-marketplace/redhat-operators-48q6k" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.753163 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99a57330-6112-4e39-aec6-3f62ff4a56ea-utilities\") pod \"redhat-operators-48q6k\" (UID: \"99a57330-6112-4e39-aec6-3f62ff4a56ea\") " pod="openshift-marketplace/redhat-operators-48q6k" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.753323 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99a57330-6112-4e39-aec6-3f62ff4a56ea-catalog-content\") pod \"redhat-operators-48q6k\" (UID: \"99a57330-6112-4e39-aec6-3f62ff4a56ea\") " pod="openshift-marketplace/redhat-operators-48q6k" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.854902 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99a57330-6112-4e39-aec6-3f62ff4a56ea-catalog-content\") pod \"redhat-operators-48q6k\" (UID: \"99a57330-6112-4e39-aec6-3f62ff4a56ea\") " pod="openshift-marketplace/redhat-operators-48q6k" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.855078 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ng4v\" (UniqueName: \"kubernetes.io/projected/99a57330-6112-4e39-aec6-3f62ff4a56ea-kube-api-access-6ng4v\") pod \"redhat-operators-48q6k\" (UID: \"99a57330-6112-4e39-aec6-3f62ff4a56ea\") " pod="openshift-marketplace/redhat-operators-48q6k" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.855115 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99a57330-6112-4e39-aec6-3f62ff4a56ea-utilities\") pod \"redhat-operators-48q6k\" (UID: \"99a57330-6112-4e39-aec6-3f62ff4a56ea\") " pod="openshift-marketplace/redhat-operators-48q6k" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.855707 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99a57330-6112-4e39-aec6-3f62ff4a56ea-catalog-content\") pod \"redhat-operators-48q6k\" (UID: \"99a57330-6112-4e39-aec6-3f62ff4a56ea\") " pod="openshift-marketplace/redhat-operators-48q6k" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.855867 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99a57330-6112-4e39-aec6-3f62ff4a56ea-utilities\") pod \"redhat-operators-48q6k\" (UID: \"99a57330-6112-4e39-aec6-3f62ff4a56ea\") " pod="openshift-marketplace/redhat-operators-48q6k" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.883614 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ng4v\" (UniqueName: \"kubernetes.io/projected/99a57330-6112-4e39-aec6-3f62ff4a56ea-kube-api-access-6ng4v\") pod \"redhat-operators-48q6k\" (UID: \"99a57330-6112-4e39-aec6-3f62ff4a56ea\") " pod="openshift-marketplace/redhat-operators-48q6k" Mar 11 13:28:17 crc kubenswrapper[4923]: I0311 13:28:17.911513 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-48q6k" Mar 11 13:28:18 crc kubenswrapper[4923]: I0311 13:28:18.153603 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-48q6k"] Mar 11 13:28:18 crc kubenswrapper[4923]: I0311 13:28:18.426224 4923 generic.go:334] "Generic (PLEG): container finished" podID="99a57330-6112-4e39-aec6-3f62ff4a56ea" containerID="b214e118b8cd91e196f5a73a7d97e66c1f2052695b9c00ae37672b324fdabf11" exitCode=0 Mar 11 13:28:18 crc kubenswrapper[4923]: I0311 13:28:18.426278 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48q6k" event={"ID":"99a57330-6112-4e39-aec6-3f62ff4a56ea","Type":"ContainerDied","Data":"b214e118b8cd91e196f5a73a7d97e66c1f2052695b9c00ae37672b324fdabf11"} Mar 11 13:28:18 crc kubenswrapper[4923]: I0311 13:28:18.426324 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48q6k" event={"ID":"99a57330-6112-4e39-aec6-3f62ff4a56ea","Type":"ContainerStarted","Data":"e1184d22c1a889da4ce74c3fe557c12d8948ed867791a58e227863ac8fea182e"} Mar 11 13:28:18 crc kubenswrapper[4923]: I0311 13:28:18.981586 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zsbfg"] Mar 11 13:28:18 crc kubenswrapper[4923]: I0311 13:28:18.983090 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zsbfg" Mar 11 13:28:18 crc kubenswrapper[4923]: I0311 13:28:18.984979 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 11 13:28:18 crc kubenswrapper[4923]: I0311 13:28:18.994483 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zsbfg"] Mar 11 13:28:19 crc kubenswrapper[4923]: I0311 13:28:19.072413 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzwv6\" (UniqueName: \"kubernetes.io/projected/38a8794f-af70-4f93-a892-13e5e6ccae32-kube-api-access-pzwv6\") pod \"certified-operators-zsbfg\" (UID: \"38a8794f-af70-4f93-a892-13e5e6ccae32\") " pod="openshift-marketplace/certified-operators-zsbfg" Mar 11 13:28:19 crc kubenswrapper[4923]: I0311 13:28:19.072479 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38a8794f-af70-4f93-a892-13e5e6ccae32-utilities\") pod \"certified-operators-zsbfg\" (UID: \"38a8794f-af70-4f93-a892-13e5e6ccae32\") " pod="openshift-marketplace/certified-operators-zsbfg" Mar 11 13:28:19 crc kubenswrapper[4923]: I0311 13:28:19.072662 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38a8794f-af70-4f93-a892-13e5e6ccae32-catalog-content\") pod \"certified-operators-zsbfg\" (UID: \"38a8794f-af70-4f93-a892-13e5e6ccae32\") " pod="openshift-marketplace/certified-operators-zsbfg" Mar 11 13:28:19 crc kubenswrapper[4923]: I0311 13:28:19.174087 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzwv6\" (UniqueName: \"kubernetes.io/projected/38a8794f-af70-4f93-a892-13e5e6ccae32-kube-api-access-pzwv6\") pod \"certified-operators-zsbfg\" (UID: \"38a8794f-af70-4f93-a892-13e5e6ccae32\") " pod="openshift-marketplace/certified-operators-zsbfg" Mar 11 13:28:19 crc kubenswrapper[4923]: I0311 13:28:19.174228 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38a8794f-af70-4f93-a892-13e5e6ccae32-utilities\") pod \"certified-operators-zsbfg\" (UID: \"38a8794f-af70-4f93-a892-13e5e6ccae32\") " pod="openshift-marketplace/certified-operators-zsbfg" Mar 11 13:28:19 crc kubenswrapper[4923]: I0311 13:28:19.174270 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38a8794f-af70-4f93-a892-13e5e6ccae32-catalog-content\") pod \"certified-operators-zsbfg\" (UID: \"38a8794f-af70-4f93-a892-13e5e6ccae32\") " pod="openshift-marketplace/certified-operators-zsbfg" Mar 11 13:28:19 crc kubenswrapper[4923]: I0311 13:28:19.175145 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38a8794f-af70-4f93-a892-13e5e6ccae32-catalog-content\") pod \"certified-operators-zsbfg\" (UID: \"38a8794f-af70-4f93-a892-13e5e6ccae32\") " pod="openshift-marketplace/certified-operators-zsbfg" Mar 11 13:28:19 crc kubenswrapper[4923]: I0311 13:28:19.175620 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38a8794f-af70-4f93-a892-13e5e6ccae32-utilities\") pod \"certified-operators-zsbfg\" (UID: \"38a8794f-af70-4f93-a892-13e5e6ccae32\") " pod="openshift-marketplace/certified-operators-zsbfg" Mar 11 13:28:19 crc kubenswrapper[4923]: I0311 13:28:19.209271 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzwv6\" (UniqueName: \"kubernetes.io/projected/38a8794f-af70-4f93-a892-13e5e6ccae32-kube-api-access-pzwv6\") pod \"certified-operators-zsbfg\" (UID: \"38a8794f-af70-4f93-a892-13e5e6ccae32\") " pod="openshift-marketplace/certified-operators-zsbfg" Mar 11 13:28:19 crc kubenswrapper[4923]: I0311 13:28:19.306532 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zsbfg" Mar 11 13:28:19 crc kubenswrapper[4923]: I0311 13:28:19.533549 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zsbfg"] Mar 11 13:28:19 crc kubenswrapper[4923]: W0311 13:28:19.536714 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod38a8794f_af70_4f93_a892_13e5e6ccae32.slice/crio-66cf06121ff346aa14d9d4455314ce2d982b3d90e2d9a4af02d6b2911593f6fb WatchSource:0}: Error finding container 66cf06121ff346aa14d9d4455314ce2d982b3d90e2d9a4af02d6b2911593f6fb: Status 404 returned error can't find the container with id 66cf06121ff346aa14d9d4455314ce2d982b3d90e2d9a4af02d6b2911593f6fb Mar 11 13:28:19 crc kubenswrapper[4923]: I0311 13:28:19.982381 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-j8cwt"] Mar 11 13:28:19 crc kubenswrapper[4923]: I0311 13:28:19.983630 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j8cwt" Mar 11 13:28:19 crc kubenswrapper[4923]: I0311 13:28:19.985825 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 11 13:28:20 crc kubenswrapper[4923]: I0311 13:28:20.009095 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j8cwt"] Mar 11 13:28:20 crc kubenswrapper[4923]: I0311 13:28:20.084031 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9-catalog-content\") pod \"community-operators-j8cwt\" (UID: \"5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9\") " pod="openshift-marketplace/community-operators-j8cwt" Mar 11 13:28:20 crc kubenswrapper[4923]: I0311 13:28:20.084135 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5m52\" (UniqueName: \"kubernetes.io/projected/5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9-kube-api-access-l5m52\") pod \"community-operators-j8cwt\" (UID: \"5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9\") " pod="openshift-marketplace/community-operators-j8cwt" Mar 11 13:28:20 crc kubenswrapper[4923]: I0311 13:28:20.084160 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9-utilities\") pod \"community-operators-j8cwt\" (UID: \"5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9\") " pod="openshift-marketplace/community-operators-j8cwt" Mar 11 13:28:20 crc kubenswrapper[4923]: I0311 13:28:20.185945 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5m52\" (UniqueName: \"kubernetes.io/projected/5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9-kube-api-access-l5m52\") pod \"community-operators-j8cwt\" (UID: \"5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9\") " pod="openshift-marketplace/community-operators-j8cwt" Mar 11 13:28:20 crc kubenswrapper[4923]: I0311 13:28:20.186394 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9-utilities\") pod \"community-operators-j8cwt\" (UID: \"5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9\") " pod="openshift-marketplace/community-operators-j8cwt" Mar 11 13:28:20 crc kubenswrapper[4923]: I0311 13:28:20.186987 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9-utilities\") pod \"community-operators-j8cwt\" (UID: \"5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9\") " pod="openshift-marketplace/community-operators-j8cwt" Mar 11 13:28:20 crc kubenswrapper[4923]: I0311 13:28:20.187131 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9-catalog-content\") pod \"community-operators-j8cwt\" (UID: \"5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9\") " pod="openshift-marketplace/community-operators-j8cwt" Mar 11 13:28:20 crc kubenswrapper[4923]: I0311 13:28:20.187477 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9-catalog-content\") pod \"community-operators-j8cwt\" (UID: \"5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9\") " pod="openshift-marketplace/community-operators-j8cwt" Mar 11 13:28:20 crc kubenswrapper[4923]: I0311 13:28:20.207523 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5m52\" (UniqueName: \"kubernetes.io/projected/5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9-kube-api-access-l5m52\") pod \"community-operators-j8cwt\" (UID: \"5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9\") " pod="openshift-marketplace/community-operators-j8cwt" Mar 11 13:28:20 crc kubenswrapper[4923]: I0311 13:28:20.296331 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j8cwt" Mar 11 13:28:20 crc kubenswrapper[4923]: I0311 13:28:20.437923 4923 generic.go:334] "Generic (PLEG): container finished" podID="38a8794f-af70-4f93-a892-13e5e6ccae32" containerID="8889e3d9f378381740cd3d4ed4d7735aa4cea07bb9093ec2313ef21c4159d941" exitCode=0 Mar 11 13:28:20 crc kubenswrapper[4923]: I0311 13:28:20.438154 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zsbfg" event={"ID":"38a8794f-af70-4f93-a892-13e5e6ccae32","Type":"ContainerDied","Data":"8889e3d9f378381740cd3d4ed4d7735aa4cea07bb9093ec2313ef21c4159d941"} Mar 11 13:28:20 crc kubenswrapper[4923]: I0311 13:28:20.438178 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zsbfg" event={"ID":"38a8794f-af70-4f93-a892-13e5e6ccae32","Type":"ContainerStarted","Data":"66cf06121ff346aa14d9d4455314ce2d982b3d90e2d9a4af02d6b2911593f6fb"} Mar 11 13:28:20 crc kubenswrapper[4923]: I0311 13:28:20.440268 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48q6k" event={"ID":"99a57330-6112-4e39-aec6-3f62ff4a56ea","Type":"ContainerStarted","Data":"6f29922a0d22388b541ff42e10ea1bcf93aa2dd5a8793e2dbf0775d89630b7d8"} Mar 11 13:28:20 crc kubenswrapper[4923]: I0311 13:28:20.587713 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j8cwt"] Mar 11 13:28:20 crc kubenswrapper[4923]: W0311 13:28:20.592286 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5986b4bc_9bbc_43a2_ab9b_2bcf145c7fb9.slice/crio-9ec776453610a8f4b12a72bb12b1e659c0f9917fceebd57a43d4af9e9bdad244 WatchSource:0}: Error finding container 9ec776453610a8f4b12a72bb12b1e659c0f9917fceebd57a43d4af9e9bdad244: Status 404 returned error can't find the container with id 9ec776453610a8f4b12a72bb12b1e659c0f9917fceebd57a43d4af9e9bdad244 Mar 11 13:28:21 crc kubenswrapper[4923]: I0311 13:28:21.390009 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x98sf"] Mar 11 13:28:21 crc kubenswrapper[4923]: I0311 13:28:21.391442 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x98sf" Mar 11 13:28:21 crc kubenswrapper[4923]: I0311 13:28:21.394376 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 11 13:28:21 crc kubenswrapper[4923]: I0311 13:28:21.409952 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x98sf"] Mar 11 13:28:21 crc kubenswrapper[4923]: I0311 13:28:21.447568 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zsbfg" event={"ID":"38a8794f-af70-4f93-a892-13e5e6ccae32","Type":"ContainerStarted","Data":"d23e1286b147f856ce875db4ab8aebfe2b7bd2cb6869fa6d54fceed454969fee"} Mar 11 13:28:21 crc kubenswrapper[4923]: I0311 13:28:21.451064 4923 generic.go:334] "Generic (PLEG): container finished" podID="99a57330-6112-4e39-aec6-3f62ff4a56ea" containerID="6f29922a0d22388b541ff42e10ea1bcf93aa2dd5a8793e2dbf0775d89630b7d8" exitCode=0 Mar 11 13:28:21 crc kubenswrapper[4923]: I0311 13:28:21.451140 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48q6k" event={"ID":"99a57330-6112-4e39-aec6-3f62ff4a56ea","Type":"ContainerDied","Data":"6f29922a0d22388b541ff42e10ea1bcf93aa2dd5a8793e2dbf0775d89630b7d8"} Mar 11 13:28:21 crc kubenswrapper[4923]: I0311 13:28:21.454202 4923 generic.go:334] "Generic (PLEG): container finished" podID="5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9" containerID="72d012fa2cf0854c8e2639bf19648aeceb31acc655bfad57f330e9a58cd24261" exitCode=0 Mar 11 13:28:21 crc kubenswrapper[4923]: I0311 13:28:21.454230 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j8cwt" event={"ID":"5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9","Type":"ContainerDied","Data":"72d012fa2cf0854c8e2639bf19648aeceb31acc655bfad57f330e9a58cd24261"} Mar 11 13:28:21 crc kubenswrapper[4923]: I0311 13:28:21.454251 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j8cwt" event={"ID":"5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9","Type":"ContainerStarted","Data":"9ec776453610a8f4b12a72bb12b1e659c0f9917fceebd57a43d4af9e9bdad244"} Mar 11 13:28:21 crc kubenswrapper[4923]: I0311 13:28:21.507544 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wv9p\" (UniqueName: \"kubernetes.io/projected/fe7957fa-3f27-4325-8e3b-5593c750bb6b-kube-api-access-5wv9p\") pod \"redhat-marketplace-x98sf\" (UID: \"fe7957fa-3f27-4325-8e3b-5593c750bb6b\") " pod="openshift-marketplace/redhat-marketplace-x98sf" Mar 11 13:28:21 crc kubenswrapper[4923]: I0311 13:28:21.507656 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe7957fa-3f27-4325-8e3b-5593c750bb6b-utilities\") pod \"redhat-marketplace-x98sf\" (UID: \"fe7957fa-3f27-4325-8e3b-5593c750bb6b\") " pod="openshift-marketplace/redhat-marketplace-x98sf" Mar 11 13:28:21 crc kubenswrapper[4923]: I0311 13:28:21.507707 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe7957fa-3f27-4325-8e3b-5593c750bb6b-catalog-content\") pod \"redhat-marketplace-x98sf\" (UID: \"fe7957fa-3f27-4325-8e3b-5593c750bb6b\") " pod="openshift-marketplace/redhat-marketplace-x98sf" Mar 11 13:28:21 crc kubenswrapper[4923]: I0311 13:28:21.609361 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wv9p\" (UniqueName: \"kubernetes.io/projected/fe7957fa-3f27-4325-8e3b-5593c750bb6b-kube-api-access-5wv9p\") pod \"redhat-marketplace-x98sf\" (UID: \"fe7957fa-3f27-4325-8e3b-5593c750bb6b\") " pod="openshift-marketplace/redhat-marketplace-x98sf" Mar 11 13:28:21 crc kubenswrapper[4923]: I0311 13:28:21.609433 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe7957fa-3f27-4325-8e3b-5593c750bb6b-utilities\") pod \"redhat-marketplace-x98sf\" (UID: \"fe7957fa-3f27-4325-8e3b-5593c750bb6b\") " pod="openshift-marketplace/redhat-marketplace-x98sf" Mar 11 13:28:21 crc kubenswrapper[4923]: I0311 13:28:21.609463 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe7957fa-3f27-4325-8e3b-5593c750bb6b-catalog-content\") pod \"redhat-marketplace-x98sf\" (UID: \"fe7957fa-3f27-4325-8e3b-5593c750bb6b\") " pod="openshift-marketplace/redhat-marketplace-x98sf" Mar 11 13:28:21 crc kubenswrapper[4923]: I0311 13:28:21.610007 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe7957fa-3f27-4325-8e3b-5593c750bb6b-catalog-content\") pod \"redhat-marketplace-x98sf\" (UID: \"fe7957fa-3f27-4325-8e3b-5593c750bb6b\") " pod="openshift-marketplace/redhat-marketplace-x98sf" Mar 11 13:28:21 crc kubenswrapper[4923]: I0311 13:28:21.610059 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe7957fa-3f27-4325-8e3b-5593c750bb6b-utilities\") pod \"redhat-marketplace-x98sf\" (UID: \"fe7957fa-3f27-4325-8e3b-5593c750bb6b\") " pod="openshift-marketplace/redhat-marketplace-x98sf" Mar 11 13:28:21 crc kubenswrapper[4923]: I0311 13:28:21.628560 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wv9p\" (UniqueName: \"kubernetes.io/projected/fe7957fa-3f27-4325-8e3b-5593c750bb6b-kube-api-access-5wv9p\") pod \"redhat-marketplace-x98sf\" (UID: \"fe7957fa-3f27-4325-8e3b-5593c750bb6b\") " pod="openshift-marketplace/redhat-marketplace-x98sf" Mar 11 13:28:21 crc kubenswrapper[4923]: I0311 13:28:21.712867 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x98sf" Mar 11 13:28:21 crc kubenswrapper[4923]: I0311 13:28:21.954944 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x98sf"] Mar 11 13:28:21 crc kubenswrapper[4923]: W0311 13:28:21.969298 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfe7957fa_3f27_4325_8e3b_5593c750bb6b.slice/crio-1ea7b84e2f0b323c66e7651d5efe612ff87c4663e350037792eebb15bf7541b3 WatchSource:0}: Error finding container 1ea7b84e2f0b323c66e7651d5efe612ff87c4663e350037792eebb15bf7541b3: Status 404 returned error can't find the container with id 1ea7b84e2f0b323c66e7651d5efe612ff87c4663e350037792eebb15bf7541b3 Mar 11 13:28:22 crc kubenswrapper[4923]: I0311 13:28:22.463382 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-48q6k" event={"ID":"99a57330-6112-4e39-aec6-3f62ff4a56ea","Type":"ContainerStarted","Data":"d46ad0c578939bfb1bf65f4e43faf3751995fd18dfef86f99a17bc64a2324d4d"} Mar 11 13:28:22 crc kubenswrapper[4923]: I0311 13:28:22.465734 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j8cwt" event={"ID":"5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9","Type":"ContainerStarted","Data":"4da3604d7754ecedca0d7deffe9801823da9671365934a83e12b102292478679"} Mar 11 13:28:22 crc kubenswrapper[4923]: I0311 13:28:22.467460 4923 generic.go:334] "Generic (PLEG): container finished" podID="fe7957fa-3f27-4325-8e3b-5593c750bb6b" containerID="14f1b2116797d9ea4c6980664528c945435d8c7873eda254cb9a4718374d8267" exitCode=0 Mar 11 13:28:22 crc kubenswrapper[4923]: I0311 13:28:22.467549 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x98sf" event={"ID":"fe7957fa-3f27-4325-8e3b-5593c750bb6b","Type":"ContainerDied","Data":"14f1b2116797d9ea4c6980664528c945435d8c7873eda254cb9a4718374d8267"} Mar 11 13:28:22 crc kubenswrapper[4923]: I0311 13:28:22.467577 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x98sf" event={"ID":"fe7957fa-3f27-4325-8e3b-5593c750bb6b","Type":"ContainerStarted","Data":"1ea7b84e2f0b323c66e7651d5efe612ff87c4663e350037792eebb15bf7541b3"} Mar 11 13:28:22 crc kubenswrapper[4923]: I0311 13:28:22.474284 4923 generic.go:334] "Generic (PLEG): container finished" podID="38a8794f-af70-4f93-a892-13e5e6ccae32" containerID="d23e1286b147f856ce875db4ab8aebfe2b7bd2cb6869fa6d54fceed454969fee" exitCode=0 Mar 11 13:28:22 crc kubenswrapper[4923]: I0311 13:28:22.474358 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zsbfg" event={"ID":"38a8794f-af70-4f93-a892-13e5e6ccae32","Type":"ContainerDied","Data":"d23e1286b147f856ce875db4ab8aebfe2b7bd2cb6869fa6d54fceed454969fee"} Mar 11 13:28:22 crc kubenswrapper[4923]: I0311 13:28:22.474391 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zsbfg" event={"ID":"38a8794f-af70-4f93-a892-13e5e6ccae32","Type":"ContainerStarted","Data":"60b51a470534bb127bdd6ccc34e2b5451cadf086b24656f148caefbbab8d224f"} Mar 11 13:28:22 crc kubenswrapper[4923]: I0311 13:28:22.497608 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-48q6k" podStartSLOduration=2.026727941 podStartE2EDuration="5.497590876s" podCreationTimestamp="2026-03-11 13:28:17 +0000 UTC" firstStartedPulling="2026-03-11 13:28:18.429949157 +0000 UTC m=+346.111850911" lastFinishedPulling="2026-03-11 13:28:21.900812102 +0000 UTC m=+349.582713846" observedRunningTime="2026-03-11 13:28:22.495695355 +0000 UTC m=+350.177597079" watchObservedRunningTime="2026-03-11 13:28:22.497590876 +0000 UTC m=+350.179492590" Mar 11 13:28:22 crc kubenswrapper[4923]: I0311 13:28:22.564132 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zsbfg" podStartSLOduration=3.008606491 podStartE2EDuration="4.564113458s" podCreationTimestamp="2026-03-11 13:28:18 +0000 UTC" firstStartedPulling="2026-03-11 13:28:20.439310736 +0000 UTC m=+348.121212450" lastFinishedPulling="2026-03-11 13:28:21.994817713 +0000 UTC m=+349.676719417" observedRunningTime="2026-03-11 13:28:22.557215081 +0000 UTC m=+350.239116835" watchObservedRunningTime="2026-03-11 13:28:22.564113458 +0000 UTC m=+350.246015182" Mar 11 13:28:23 crc kubenswrapper[4923]: I0311 13:28:23.482265 4923 generic.go:334] "Generic (PLEG): container finished" podID="5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9" containerID="4da3604d7754ecedca0d7deffe9801823da9671365934a83e12b102292478679" exitCode=0 Mar 11 13:28:23 crc kubenswrapper[4923]: I0311 13:28:23.482320 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j8cwt" event={"ID":"5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9","Type":"ContainerDied","Data":"4da3604d7754ecedca0d7deffe9801823da9671365934a83e12b102292478679"} Mar 11 13:28:23 crc kubenswrapper[4923]: I0311 13:28:23.490577 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x98sf" event={"ID":"fe7957fa-3f27-4325-8e3b-5593c750bb6b","Type":"ContainerStarted","Data":"d4ea01a1e8bc33a3c6cddd1dc3deb5ba6974c979d49eb325de0af839ac232522"} Mar 11 13:28:24 crc kubenswrapper[4923]: I0311 13:28:24.498180 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j8cwt" event={"ID":"5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9","Type":"ContainerStarted","Data":"5f70cebf83a639c02394a2c9f4140b5194eb85a07fab58058cbb9bf2641c235e"} Mar 11 13:28:24 crc kubenswrapper[4923]: I0311 13:28:24.499846 4923 generic.go:334] "Generic (PLEG): container finished" podID="fe7957fa-3f27-4325-8e3b-5593c750bb6b" containerID="d4ea01a1e8bc33a3c6cddd1dc3deb5ba6974c979d49eb325de0af839ac232522" exitCode=0 Mar 11 13:28:24 crc kubenswrapper[4923]: I0311 13:28:24.499906 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x98sf" event={"ID":"fe7957fa-3f27-4325-8e3b-5593c750bb6b","Type":"ContainerDied","Data":"d4ea01a1e8bc33a3c6cddd1dc3deb5ba6974c979d49eb325de0af839ac232522"} Mar 11 13:28:24 crc kubenswrapper[4923]: I0311 13:28:24.524919 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-j8cwt" podStartSLOduration=3.120195388 podStartE2EDuration="5.524900764s" podCreationTimestamp="2026-03-11 13:28:19 +0000 UTC" firstStartedPulling="2026-03-11 13:28:21.458044413 +0000 UTC m=+349.139946117" lastFinishedPulling="2026-03-11 13:28:23.862749749 +0000 UTC m=+351.544651493" observedRunningTime="2026-03-11 13:28:24.522264132 +0000 UTC m=+352.204165866" watchObservedRunningTime="2026-03-11 13:28:24.524900764 +0000 UTC m=+352.206802478" Mar 11 13:28:25 crc kubenswrapper[4923]: I0311 13:28:25.529367 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x98sf" event={"ID":"fe7957fa-3f27-4325-8e3b-5593c750bb6b","Type":"ContainerStarted","Data":"ef1d60c3d283e7d7c1168025c679fbb88770860ed54dd871d58811fce993b8f2"} Mar 11 13:28:25 crc kubenswrapper[4923]: I0311 13:28:25.556119 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x98sf" podStartSLOduration=2.058127854 podStartE2EDuration="4.55610492s" podCreationTimestamp="2026-03-11 13:28:21 +0000 UTC" firstStartedPulling="2026-03-11 13:28:22.469705617 +0000 UTC m=+350.151607331" lastFinishedPulling="2026-03-11 13:28:24.967682673 +0000 UTC m=+352.649584397" observedRunningTime="2026-03-11 13:28:25.552066781 +0000 UTC m=+353.233968495" watchObservedRunningTime="2026-03-11 13:28:25.55610492 +0000 UTC m=+353.238006634" Mar 11 13:28:27 crc kubenswrapper[4923]: I0311 13:28:27.912621 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-48q6k" Mar 11 13:28:27 crc kubenswrapper[4923]: I0311 13:28:27.912780 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-48q6k" Mar 11 13:28:28 crc kubenswrapper[4923]: I0311 13:28:28.973980 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-48q6k" podUID="99a57330-6112-4e39-aec6-3f62ff4a56ea" containerName="registry-server" probeResult="failure" output=< Mar 11 13:28:28 crc kubenswrapper[4923]: timeout: failed to connect service ":50051" within 1s Mar 11 13:28:28 crc kubenswrapper[4923]: > Mar 11 13:28:29 crc kubenswrapper[4923]: I0311 13:28:29.306997 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zsbfg" Mar 11 13:28:29 crc kubenswrapper[4923]: I0311 13:28:29.309312 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zsbfg" Mar 11 13:28:29 crc kubenswrapper[4923]: I0311 13:28:29.360273 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zsbfg" Mar 11 13:28:29 crc kubenswrapper[4923]: I0311 13:28:29.598383 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zsbfg" Mar 11 13:28:30 crc kubenswrapper[4923]: I0311 13:28:30.297505 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-j8cwt" Mar 11 13:28:30 crc kubenswrapper[4923]: I0311 13:28:30.297596 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-j8cwt" Mar 11 13:28:30 crc kubenswrapper[4923]: I0311 13:28:30.369460 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-j8cwt" Mar 11 13:28:30 crc kubenswrapper[4923]: I0311 13:28:30.636364 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-j8cwt" Mar 11 13:28:31 crc kubenswrapper[4923]: I0311 13:28:31.714113 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x98sf" Mar 11 13:28:31 crc kubenswrapper[4923]: I0311 13:28:31.714154 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x98sf" Mar 11 13:28:31 crc kubenswrapper[4923]: I0311 13:28:31.769941 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x98sf" Mar 11 13:28:32 crc kubenswrapper[4923]: I0311 13:28:32.640572 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x98sf" Mar 11 13:28:33 crc kubenswrapper[4923]: I0311 13:28:33.585705 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-5zbtq" Mar 11 13:28:33 crc kubenswrapper[4923]: I0311 13:28:33.675789 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fwfgr"] Mar 11 13:28:37 crc kubenswrapper[4923]: I0311 13:28:37.989035 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-48q6k" Mar 11 13:28:38 crc kubenswrapper[4923]: I0311 13:28:38.060182 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-48q6k" Mar 11 13:28:58 crc kubenswrapper[4923]: I0311 13:28:58.729630 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" podUID="b819b4e5-85bd-4a31-bc2e-a119da615675" containerName="registry" containerID="cri-o://924e8bb4b0d1749a95af14ab43eb2838fb04a0aeab2b3088e568931d3a87ec67" gracePeriod=30 Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.175222 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.344229 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b819b4e5-85bd-4a31-bc2e-a119da615675-registry-certificates\") pod \"b819b4e5-85bd-4a31-bc2e-a119da615675\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.344312 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8k8rp\" (UniqueName: \"kubernetes.io/projected/b819b4e5-85bd-4a31-bc2e-a119da615675-kube-api-access-8k8rp\") pod \"b819b4e5-85bd-4a31-bc2e-a119da615675\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.344352 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b819b4e5-85bd-4a31-bc2e-a119da615675-ca-trust-extracted\") pod \"b819b4e5-85bd-4a31-bc2e-a119da615675\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.345395 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b819b4e5-85bd-4a31-bc2e-a119da615675-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "b819b4e5-85bd-4a31-bc2e-a119da615675" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.346531 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"b819b4e5-85bd-4a31-bc2e-a119da615675\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.346647 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b819b4e5-85bd-4a31-bc2e-a119da615675-trusted-ca\") pod \"b819b4e5-85bd-4a31-bc2e-a119da615675\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.346681 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b819b4e5-85bd-4a31-bc2e-a119da615675-bound-sa-token\") pod \"b819b4e5-85bd-4a31-bc2e-a119da615675\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.346713 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b819b4e5-85bd-4a31-bc2e-a119da615675-installation-pull-secrets\") pod \"b819b4e5-85bd-4a31-bc2e-a119da615675\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.346734 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b819b4e5-85bd-4a31-bc2e-a119da615675-registry-tls\") pod \"b819b4e5-85bd-4a31-bc2e-a119da615675\" (UID: \"b819b4e5-85bd-4a31-bc2e-a119da615675\") " Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.347088 4923 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/b819b4e5-85bd-4a31-bc2e-a119da615675-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.347436 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b819b4e5-85bd-4a31-bc2e-a119da615675-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "b819b4e5-85bd-4a31-bc2e-a119da615675" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.353775 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b819b4e5-85bd-4a31-bc2e-a119da615675-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "b819b4e5-85bd-4a31-bc2e-a119da615675" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.355681 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b819b4e5-85bd-4a31-bc2e-a119da615675-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "b819b4e5-85bd-4a31-bc2e-a119da615675" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.355973 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b819b4e5-85bd-4a31-bc2e-a119da615675-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "b819b4e5-85bd-4a31-bc2e-a119da615675" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.358459 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b819b4e5-85bd-4a31-bc2e-a119da615675-kube-api-access-8k8rp" (OuterVolumeSpecName: "kube-api-access-8k8rp") pod "b819b4e5-85bd-4a31-bc2e-a119da615675" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675"). InnerVolumeSpecName "kube-api-access-8k8rp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.359948 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b819b4e5-85bd-4a31-bc2e-a119da615675-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "b819b4e5-85bd-4a31-bc2e-a119da615675" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.365225 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "b819b4e5-85bd-4a31-bc2e-a119da615675" (UID: "b819b4e5-85bd-4a31-bc2e-a119da615675"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.448000 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8k8rp\" (UniqueName: \"kubernetes.io/projected/b819b4e5-85bd-4a31-bc2e-a119da615675-kube-api-access-8k8rp\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.448036 4923 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/b819b4e5-85bd-4a31-bc2e-a119da615675-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.448048 4923 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b819b4e5-85bd-4a31-bc2e-a119da615675-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.448058 4923 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b819b4e5-85bd-4a31-bc2e-a119da615675-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.448070 4923 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/b819b4e5-85bd-4a31-bc2e-a119da615675-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.448079 4923 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/b819b4e5-85bd-4a31-bc2e-a119da615675-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.781501 4923 generic.go:334] "Generic (PLEG): container finished" podID="b819b4e5-85bd-4a31-bc2e-a119da615675" containerID="924e8bb4b0d1749a95af14ab43eb2838fb04a0aeab2b3088e568931d3a87ec67" exitCode=0 Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.781552 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" event={"ID":"b819b4e5-85bd-4a31-bc2e-a119da615675","Type":"ContainerDied","Data":"924e8bb4b0d1749a95af14ab43eb2838fb04a0aeab2b3088e568931d3a87ec67"} Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.781594 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" event={"ID":"b819b4e5-85bd-4a31-bc2e-a119da615675","Type":"ContainerDied","Data":"5e152f856cc842a14bf2685d7d785ce4b1198d88a2087c5df0e7856a7ef962f6"} Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.781614 4923 scope.go:117] "RemoveContainer" containerID="924e8bb4b0d1749a95af14ab43eb2838fb04a0aeab2b3088e568931d3a87ec67" Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.781580 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fwfgr" Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.801649 4923 scope.go:117] "RemoveContainer" containerID="924e8bb4b0d1749a95af14ab43eb2838fb04a0aeab2b3088e568931d3a87ec67" Mar 11 13:28:59 crc kubenswrapper[4923]: E0311 13:28:59.803732 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"924e8bb4b0d1749a95af14ab43eb2838fb04a0aeab2b3088e568931d3a87ec67\": container with ID starting with 924e8bb4b0d1749a95af14ab43eb2838fb04a0aeab2b3088e568931d3a87ec67 not found: ID does not exist" containerID="924e8bb4b0d1749a95af14ab43eb2838fb04a0aeab2b3088e568931d3a87ec67" Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.803770 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"924e8bb4b0d1749a95af14ab43eb2838fb04a0aeab2b3088e568931d3a87ec67"} err="failed to get container status \"924e8bb4b0d1749a95af14ab43eb2838fb04a0aeab2b3088e568931d3a87ec67\": rpc error: code = NotFound desc = could not find container \"924e8bb4b0d1749a95af14ab43eb2838fb04a0aeab2b3088e568931d3a87ec67\": container with ID starting with 924e8bb4b0d1749a95af14ab43eb2838fb04a0aeab2b3088e568931d3a87ec67 not found: ID does not exist" Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.827703 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fwfgr"] Mar 11 13:28:59 crc kubenswrapper[4923]: I0311 13:28:59.839740 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fwfgr"] Mar 11 13:29:01 crc kubenswrapper[4923]: I0311 13:29:01.032260 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b819b4e5-85bd-4a31-bc2e-a119da615675" path="/var/lib/kubelet/pods/b819b4e5-85bd-4a31-bc2e-a119da615675/volumes" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.154089 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29553930-mqfkr"] Mar 11 13:30:00 crc kubenswrapper[4923]: E0311 13:30:00.154897 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b819b4e5-85bd-4a31-bc2e-a119da615675" containerName="registry" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.154914 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="b819b4e5-85bd-4a31-bc2e-a119da615675" containerName="registry" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.155023 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="b819b4e5-85bd-4a31-bc2e-a119da615675" containerName="registry" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.155462 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553930-mqfkr" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.159575 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.160113 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.162938 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-nxck7" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.168452 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29553930-7bsvh"] Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.169569 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29553930-7bsvh" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.173287 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.173419 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.183881 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29553930-7bsvh"] Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.228061 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553930-mqfkr"] Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.269870 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt5j4\" (UniqueName: \"kubernetes.io/projected/6fbadffd-f924-474e-af04-11fd967a67a6-kube-api-access-rt5j4\") pod \"auto-csr-approver-29553930-mqfkr\" (UID: \"6fbadffd-f924-474e-af04-11fd967a67a6\") " pod="openshift-infra/auto-csr-approver-29553930-mqfkr" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.269998 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6adbaa00-092b-4f5b-912c-20d638976c2e-config-volume\") pod \"collect-profiles-29553930-7bsvh\" (UID: \"6adbaa00-092b-4f5b-912c-20d638976c2e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553930-7bsvh" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.270038 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6adbaa00-092b-4f5b-912c-20d638976c2e-secret-volume\") pod \"collect-profiles-29553930-7bsvh\" (UID: \"6adbaa00-092b-4f5b-912c-20d638976c2e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553930-7bsvh" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.270115 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nmwb\" (UniqueName: \"kubernetes.io/projected/6adbaa00-092b-4f5b-912c-20d638976c2e-kube-api-access-2nmwb\") pod \"collect-profiles-29553930-7bsvh\" (UID: \"6adbaa00-092b-4f5b-912c-20d638976c2e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553930-7bsvh" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.371564 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt5j4\" (UniqueName: \"kubernetes.io/projected/6fbadffd-f924-474e-af04-11fd967a67a6-kube-api-access-rt5j4\") pod \"auto-csr-approver-29553930-mqfkr\" (UID: \"6fbadffd-f924-474e-af04-11fd967a67a6\") " pod="openshift-infra/auto-csr-approver-29553930-mqfkr" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.371649 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6adbaa00-092b-4f5b-912c-20d638976c2e-config-volume\") pod \"collect-profiles-29553930-7bsvh\" (UID: \"6adbaa00-092b-4f5b-912c-20d638976c2e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553930-7bsvh" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.371704 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6adbaa00-092b-4f5b-912c-20d638976c2e-secret-volume\") pod \"collect-profiles-29553930-7bsvh\" (UID: \"6adbaa00-092b-4f5b-912c-20d638976c2e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553930-7bsvh" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.371878 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nmwb\" (UniqueName: \"kubernetes.io/projected/6adbaa00-092b-4f5b-912c-20d638976c2e-kube-api-access-2nmwb\") pod \"collect-profiles-29553930-7bsvh\" (UID: \"6adbaa00-092b-4f5b-912c-20d638976c2e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553930-7bsvh" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.372485 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6adbaa00-092b-4f5b-912c-20d638976c2e-config-volume\") pod \"collect-profiles-29553930-7bsvh\" (UID: \"6adbaa00-092b-4f5b-912c-20d638976c2e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553930-7bsvh" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.380409 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6adbaa00-092b-4f5b-912c-20d638976c2e-secret-volume\") pod \"collect-profiles-29553930-7bsvh\" (UID: \"6adbaa00-092b-4f5b-912c-20d638976c2e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553930-7bsvh" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.388148 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt5j4\" (UniqueName: \"kubernetes.io/projected/6fbadffd-f924-474e-af04-11fd967a67a6-kube-api-access-rt5j4\") pod \"auto-csr-approver-29553930-mqfkr\" (UID: \"6fbadffd-f924-474e-af04-11fd967a67a6\") " pod="openshift-infra/auto-csr-approver-29553930-mqfkr" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.390075 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nmwb\" (UniqueName: \"kubernetes.io/projected/6adbaa00-092b-4f5b-912c-20d638976c2e-kube-api-access-2nmwb\") pod \"collect-profiles-29553930-7bsvh\" (UID: \"6adbaa00-092b-4f5b-912c-20d638976c2e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553930-7bsvh" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.477043 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553930-mqfkr" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.496779 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29553930-7bsvh" Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.726387 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553930-mqfkr"] Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.732842 4923 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 11 13:30:00 crc kubenswrapper[4923]: I0311 13:30:00.765428 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29553930-7bsvh"] Mar 11 13:30:01 crc kubenswrapper[4923]: I0311 13:30:01.212192 4923 generic.go:334] "Generic (PLEG): container finished" podID="6adbaa00-092b-4f5b-912c-20d638976c2e" containerID="93011a719f9e4ccf2bd7bc6567c5cfd1e5798984bb6367036412383326f68663" exitCode=0 Mar 11 13:30:01 crc kubenswrapper[4923]: I0311 13:30:01.212254 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29553930-7bsvh" event={"ID":"6adbaa00-092b-4f5b-912c-20d638976c2e","Type":"ContainerDied","Data":"93011a719f9e4ccf2bd7bc6567c5cfd1e5798984bb6367036412383326f68663"} Mar 11 13:30:01 crc kubenswrapper[4923]: I0311 13:30:01.212317 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29553930-7bsvh" event={"ID":"6adbaa00-092b-4f5b-912c-20d638976c2e","Type":"ContainerStarted","Data":"f09c5cf197d1ed2339008cae195d165d30f8650eb13a25c537e6001b175635e1"} Mar 11 13:30:01 crc kubenswrapper[4923]: I0311 13:30:01.213644 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553930-mqfkr" event={"ID":"6fbadffd-f924-474e-af04-11fd967a67a6","Type":"ContainerStarted","Data":"1705c659f7c5338fc734efa7cc67f3081c2bd588867a5c462fd7f8b17108fffd"} Mar 11 13:30:02 crc kubenswrapper[4923]: I0311 13:30:02.724672 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29553930-7bsvh" Mar 11 13:30:02 crc kubenswrapper[4923]: I0311 13:30:02.904839 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6adbaa00-092b-4f5b-912c-20d638976c2e-secret-volume\") pod \"6adbaa00-092b-4f5b-912c-20d638976c2e\" (UID: \"6adbaa00-092b-4f5b-912c-20d638976c2e\") " Mar 11 13:30:02 crc kubenswrapper[4923]: I0311 13:30:02.904900 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nmwb\" (UniqueName: \"kubernetes.io/projected/6adbaa00-092b-4f5b-912c-20d638976c2e-kube-api-access-2nmwb\") pod \"6adbaa00-092b-4f5b-912c-20d638976c2e\" (UID: \"6adbaa00-092b-4f5b-912c-20d638976c2e\") " Mar 11 13:30:02 crc kubenswrapper[4923]: I0311 13:30:02.904946 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6adbaa00-092b-4f5b-912c-20d638976c2e-config-volume\") pod \"6adbaa00-092b-4f5b-912c-20d638976c2e\" (UID: \"6adbaa00-092b-4f5b-912c-20d638976c2e\") " Mar 11 13:30:02 crc kubenswrapper[4923]: I0311 13:30:02.905913 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6adbaa00-092b-4f5b-912c-20d638976c2e-config-volume" (OuterVolumeSpecName: "config-volume") pod "6adbaa00-092b-4f5b-912c-20d638976c2e" (UID: "6adbaa00-092b-4f5b-912c-20d638976c2e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:30:02 crc kubenswrapper[4923]: I0311 13:30:02.914515 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6adbaa00-092b-4f5b-912c-20d638976c2e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "6adbaa00-092b-4f5b-912c-20d638976c2e" (UID: "6adbaa00-092b-4f5b-912c-20d638976c2e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:30:02 crc kubenswrapper[4923]: I0311 13:30:02.922802 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6adbaa00-092b-4f5b-912c-20d638976c2e-kube-api-access-2nmwb" (OuterVolumeSpecName: "kube-api-access-2nmwb") pod "6adbaa00-092b-4f5b-912c-20d638976c2e" (UID: "6adbaa00-092b-4f5b-912c-20d638976c2e"). InnerVolumeSpecName "kube-api-access-2nmwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:30:03 crc kubenswrapper[4923]: I0311 13:30:03.006390 4923 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/6adbaa00-092b-4f5b-912c-20d638976c2e-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 11 13:30:03 crc kubenswrapper[4923]: I0311 13:30:03.006712 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nmwb\" (UniqueName: \"kubernetes.io/projected/6adbaa00-092b-4f5b-912c-20d638976c2e-kube-api-access-2nmwb\") on node \"crc\" DevicePath \"\"" Mar 11 13:30:03 crc kubenswrapper[4923]: I0311 13:30:03.006726 4923 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/6adbaa00-092b-4f5b-912c-20d638976c2e-config-volume\") on node \"crc\" DevicePath \"\"" Mar 11 13:30:03 crc kubenswrapper[4923]: I0311 13:30:03.518210 4923 generic.go:334] "Generic (PLEG): container finished" podID="6fbadffd-f924-474e-af04-11fd967a67a6" containerID="a9274dcde3350104b71357d8a252a2cd660c94e7fa939f94b0898d87084ff6f7" exitCode=0 Mar 11 13:30:03 crc kubenswrapper[4923]: I0311 13:30:03.518276 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553930-mqfkr" event={"ID":"6fbadffd-f924-474e-af04-11fd967a67a6","Type":"ContainerDied","Data":"a9274dcde3350104b71357d8a252a2cd660c94e7fa939f94b0898d87084ff6f7"} Mar 11 13:30:03 crc kubenswrapper[4923]: I0311 13:30:03.521753 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29553930-7bsvh" event={"ID":"6adbaa00-092b-4f5b-912c-20d638976c2e","Type":"ContainerDied","Data":"f09c5cf197d1ed2339008cae195d165d30f8650eb13a25c537e6001b175635e1"} Mar 11 13:30:03 crc kubenswrapper[4923]: I0311 13:30:03.521782 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f09c5cf197d1ed2339008cae195d165d30f8650eb13a25c537e6001b175635e1" Mar 11 13:30:03 crc kubenswrapper[4923]: I0311 13:30:03.521821 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29553930-7bsvh" Mar 11 13:30:04 crc kubenswrapper[4923]: I0311 13:30:04.769654 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553930-mqfkr" Mar 11 13:30:04 crc kubenswrapper[4923]: I0311 13:30:04.946406 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rt5j4\" (UniqueName: \"kubernetes.io/projected/6fbadffd-f924-474e-af04-11fd967a67a6-kube-api-access-rt5j4\") pod \"6fbadffd-f924-474e-af04-11fd967a67a6\" (UID: \"6fbadffd-f924-474e-af04-11fd967a67a6\") " Mar 11 13:30:04 crc kubenswrapper[4923]: I0311 13:30:04.953274 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fbadffd-f924-474e-af04-11fd967a67a6-kube-api-access-rt5j4" (OuterVolumeSpecName: "kube-api-access-rt5j4") pod "6fbadffd-f924-474e-af04-11fd967a67a6" (UID: "6fbadffd-f924-474e-af04-11fd967a67a6"). InnerVolumeSpecName "kube-api-access-rt5j4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:30:05 crc kubenswrapper[4923]: I0311 13:30:05.047383 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rt5j4\" (UniqueName: \"kubernetes.io/projected/6fbadffd-f924-474e-af04-11fd967a67a6-kube-api-access-rt5j4\") on node \"crc\" DevicePath \"\"" Mar 11 13:30:05 crc kubenswrapper[4923]: I0311 13:30:05.533083 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553930-mqfkr" event={"ID":"6fbadffd-f924-474e-af04-11fd967a67a6","Type":"ContainerDied","Data":"1705c659f7c5338fc734efa7cc67f3081c2bd588867a5c462fd7f8b17108fffd"} Mar 11 13:30:05 crc kubenswrapper[4923]: I0311 13:30:05.533118 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1705c659f7c5338fc734efa7cc67f3081c2bd588867a5c462fd7f8b17108fffd" Mar 11 13:30:05 crc kubenswrapper[4923]: I0311 13:30:05.533160 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553930-mqfkr" Mar 11 13:30:11 crc kubenswrapper[4923]: I0311 13:30:11.932798 4923 patch_prober.go:28] interesting pod/machine-config-daemon-fsfnw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 11 13:30:11 crc kubenswrapper[4923]: I0311 13:30:11.933575 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 11 13:30:41 crc kubenswrapper[4923]: I0311 13:30:41.932973 4923 patch_prober.go:28] interesting pod/machine-config-daemon-fsfnw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 11 13:30:41 crc kubenswrapper[4923]: I0311 13:30:41.933700 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 11 13:31:11 crc kubenswrapper[4923]: I0311 13:31:11.932946 4923 patch_prober.go:28] interesting pod/machine-config-daemon-fsfnw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 11 13:31:11 crc kubenswrapper[4923]: I0311 13:31:11.933755 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 11 13:31:11 crc kubenswrapper[4923]: I0311 13:31:11.933844 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" Mar 11 13:31:11 crc kubenswrapper[4923]: I0311 13:31:11.934922 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e24d6b8a7efc75b0eaec16ef2d2e5b49a3a73ee933ce1436580a5243233d6453"} pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 11 13:31:11 crc kubenswrapper[4923]: I0311 13:31:11.935026 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" containerID="cri-o://e24d6b8a7efc75b0eaec16ef2d2e5b49a3a73ee933ce1436580a5243233d6453" gracePeriod=600 Mar 11 13:31:13 crc kubenswrapper[4923]: I0311 13:31:13.013103 4923 generic.go:334] "Generic (PLEG): container finished" podID="356de301-ded3-4112-8406-1a28e44c2082" containerID="e24d6b8a7efc75b0eaec16ef2d2e5b49a3a73ee933ce1436580a5243233d6453" exitCode=0 Mar 11 13:31:13 crc kubenswrapper[4923]: I0311 13:31:13.013211 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" event={"ID":"356de301-ded3-4112-8406-1a28e44c2082","Type":"ContainerDied","Data":"e24d6b8a7efc75b0eaec16ef2d2e5b49a3a73ee933ce1436580a5243233d6453"} Mar 11 13:31:13 crc kubenswrapper[4923]: I0311 13:31:13.013902 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" event={"ID":"356de301-ded3-4112-8406-1a28e44c2082","Type":"ContainerStarted","Data":"e9d14dc9fced1f5ac360fb70b07bfe3d2a670f2c82a614a9b6443f96729aaa50"} Mar 11 13:31:13 crc kubenswrapper[4923]: I0311 13:31:13.013937 4923 scope.go:117] "RemoveContainer" containerID="e04ff4641c7ac0b2e854199edde21a639f88c1a928f8af341f17e144a3a65e5d" Mar 11 13:31:33 crc kubenswrapper[4923]: I0311 13:31:33.395612 4923 scope.go:117] "RemoveContainer" containerID="784602a924d09abfb2e1a681cacf54d66ca459bb02d50b9e2b24ba71177c485a" Mar 11 13:31:33 crc kubenswrapper[4923]: I0311 13:31:33.424392 4923 scope.go:117] "RemoveContainer" containerID="84a62c4f1b65518a0426df3f6f362d9e4629e3340435f67003dc4864fce6b196" Mar 11 13:31:33 crc kubenswrapper[4923]: I0311 13:31:33.443269 4923 scope.go:117] "RemoveContainer" containerID="c573b74eea83b701b8f005938e15c3d4ec63a4098b8bdd76d2ba418a747883cd" Mar 11 13:32:00 crc kubenswrapper[4923]: I0311 13:32:00.136858 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29553932-577nf"] Mar 11 13:32:00 crc kubenswrapper[4923]: E0311 13:32:00.137847 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fbadffd-f924-474e-af04-11fd967a67a6" containerName="oc" Mar 11 13:32:00 crc kubenswrapper[4923]: I0311 13:32:00.137866 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fbadffd-f924-474e-af04-11fd967a67a6" containerName="oc" Mar 11 13:32:00 crc kubenswrapper[4923]: E0311 13:32:00.137891 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6adbaa00-092b-4f5b-912c-20d638976c2e" containerName="collect-profiles" Mar 11 13:32:00 crc kubenswrapper[4923]: I0311 13:32:00.137900 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="6adbaa00-092b-4f5b-912c-20d638976c2e" containerName="collect-profiles" Mar 11 13:32:00 crc kubenswrapper[4923]: I0311 13:32:00.138111 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="6adbaa00-092b-4f5b-912c-20d638976c2e" containerName="collect-profiles" Mar 11 13:32:00 crc kubenswrapper[4923]: I0311 13:32:00.138135 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fbadffd-f924-474e-af04-11fd967a67a6" containerName="oc" Mar 11 13:32:00 crc kubenswrapper[4923]: I0311 13:32:00.138673 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553932-577nf" Mar 11 13:32:00 crc kubenswrapper[4923]: I0311 13:32:00.140828 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 11 13:32:00 crc kubenswrapper[4923]: I0311 13:32:00.141400 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-nxck7" Mar 11 13:32:00 crc kubenswrapper[4923]: I0311 13:32:00.143662 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 11 13:32:00 crc kubenswrapper[4923]: I0311 13:32:00.148616 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553932-577nf"] Mar 11 13:32:00 crc kubenswrapper[4923]: I0311 13:32:00.246432 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78kr9\" (UniqueName: \"kubernetes.io/projected/39d4c595-2048-4b75-9c9e-e7d7d817afd7-kube-api-access-78kr9\") pod \"auto-csr-approver-29553932-577nf\" (UID: \"39d4c595-2048-4b75-9c9e-e7d7d817afd7\") " pod="openshift-infra/auto-csr-approver-29553932-577nf" Mar 11 13:32:00 crc kubenswrapper[4923]: I0311 13:32:00.347644 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78kr9\" (UniqueName: \"kubernetes.io/projected/39d4c595-2048-4b75-9c9e-e7d7d817afd7-kube-api-access-78kr9\") pod \"auto-csr-approver-29553932-577nf\" (UID: \"39d4c595-2048-4b75-9c9e-e7d7d817afd7\") " pod="openshift-infra/auto-csr-approver-29553932-577nf" Mar 11 13:32:00 crc kubenswrapper[4923]: I0311 13:32:00.390629 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78kr9\" (UniqueName: \"kubernetes.io/projected/39d4c595-2048-4b75-9c9e-e7d7d817afd7-kube-api-access-78kr9\") pod \"auto-csr-approver-29553932-577nf\" (UID: \"39d4c595-2048-4b75-9c9e-e7d7d817afd7\") " pod="openshift-infra/auto-csr-approver-29553932-577nf" Mar 11 13:32:00 crc kubenswrapper[4923]: I0311 13:32:00.471924 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553932-577nf" Mar 11 13:32:00 crc kubenswrapper[4923]: I0311 13:32:00.719120 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553932-577nf"] Mar 11 13:32:01 crc kubenswrapper[4923]: I0311 13:32:01.342975 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553932-577nf" event={"ID":"39d4c595-2048-4b75-9c9e-e7d7d817afd7","Type":"ContainerStarted","Data":"97f7a1d7d60ddf2154c95aacd17eb293109b8973183facb701faae8f1f52236c"} Mar 11 13:32:02 crc kubenswrapper[4923]: I0311 13:32:02.370590 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553932-577nf" event={"ID":"39d4c595-2048-4b75-9c9e-e7d7d817afd7","Type":"ContainerStarted","Data":"462bd3b780876dcdd389b70651d32cd1355063a305d760959bd250f0208c97fa"} Mar 11 13:32:02 crc kubenswrapper[4923]: I0311 13:32:02.395621 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29553932-577nf" podStartSLOduration=1.255190198 podStartE2EDuration="2.395599712s" podCreationTimestamp="2026-03-11 13:32:00 +0000 UTC" firstStartedPulling="2026-03-11 13:32:00.727757339 +0000 UTC m=+568.409659053" lastFinishedPulling="2026-03-11 13:32:01.868166813 +0000 UTC m=+569.550068567" observedRunningTime="2026-03-11 13:32:02.392036555 +0000 UTC m=+570.073938279" watchObservedRunningTime="2026-03-11 13:32:02.395599712 +0000 UTC m=+570.077501436" Mar 11 13:32:03 crc kubenswrapper[4923]: I0311 13:32:03.379664 4923 generic.go:334] "Generic (PLEG): container finished" podID="39d4c595-2048-4b75-9c9e-e7d7d817afd7" containerID="462bd3b780876dcdd389b70651d32cd1355063a305d760959bd250f0208c97fa" exitCode=0 Mar 11 13:32:03 crc kubenswrapper[4923]: I0311 13:32:03.379740 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553932-577nf" event={"ID":"39d4c595-2048-4b75-9c9e-e7d7d817afd7","Type":"ContainerDied","Data":"462bd3b780876dcdd389b70651d32cd1355063a305d760959bd250f0208c97fa"} Mar 11 13:32:04 crc kubenswrapper[4923]: I0311 13:32:04.672194 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553932-577nf" Mar 11 13:32:04 crc kubenswrapper[4923]: I0311 13:32:04.805106 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78kr9\" (UniqueName: \"kubernetes.io/projected/39d4c595-2048-4b75-9c9e-e7d7d817afd7-kube-api-access-78kr9\") pod \"39d4c595-2048-4b75-9c9e-e7d7d817afd7\" (UID: \"39d4c595-2048-4b75-9c9e-e7d7d817afd7\") " Mar 11 13:32:04 crc kubenswrapper[4923]: I0311 13:32:04.814753 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39d4c595-2048-4b75-9c9e-e7d7d817afd7-kube-api-access-78kr9" (OuterVolumeSpecName: "kube-api-access-78kr9") pod "39d4c595-2048-4b75-9c9e-e7d7d817afd7" (UID: "39d4c595-2048-4b75-9c9e-e7d7d817afd7"). InnerVolumeSpecName "kube-api-access-78kr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:32:04 crc kubenswrapper[4923]: I0311 13:32:04.907102 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78kr9\" (UniqueName: \"kubernetes.io/projected/39d4c595-2048-4b75-9c9e-e7d7d817afd7-kube-api-access-78kr9\") on node \"crc\" DevicePath \"\"" Mar 11 13:32:05 crc kubenswrapper[4923]: I0311 13:32:05.397254 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553932-577nf" event={"ID":"39d4c595-2048-4b75-9c9e-e7d7d817afd7","Type":"ContainerDied","Data":"97f7a1d7d60ddf2154c95aacd17eb293109b8973183facb701faae8f1f52236c"} Mar 11 13:32:05 crc kubenswrapper[4923]: I0311 13:32:05.397313 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97f7a1d7d60ddf2154c95aacd17eb293109b8973183facb701faae8f1f52236c" Mar 11 13:32:05 crc kubenswrapper[4923]: I0311 13:32:05.397424 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553932-577nf" Mar 11 13:32:05 crc kubenswrapper[4923]: I0311 13:32:05.466973 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29553926-tm54x"] Mar 11 13:32:05 crc kubenswrapper[4923]: I0311 13:32:05.467692 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29553926-tm54x"] Mar 11 13:32:07 crc kubenswrapper[4923]: I0311 13:32:07.045759 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efe004c2-3d9c-4ccb-b6c8-fbc5aa9dd615" path="/var/lib/kubelet/pods/efe004c2-3d9c-4ccb-b6c8-fbc5aa9dd615/volumes" Mar 11 13:33:33 crc kubenswrapper[4923]: I0311 13:33:33.511150 4923 scope.go:117] "RemoveContainer" containerID="cba92cc62a228910fdcadcc1532b801949cd6ea350cd346bd309e4fb4b3569da" Mar 11 13:33:41 crc kubenswrapper[4923]: I0311 13:33:41.932319 4923 patch_prober.go:28] interesting pod/machine-config-daemon-fsfnw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 11 13:33:41 crc kubenswrapper[4923]: I0311 13:33:41.932940 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 11 13:33:53 crc kubenswrapper[4923]: I0311 13:33:53.656076 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-552rb"] Mar 11 13:33:53 crc kubenswrapper[4923]: I0311 13:33:53.657247 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="ovn-controller" containerID="cri-o://6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595" gracePeriod=30 Mar 11 13:33:53 crc kubenswrapper[4923]: I0311 13:33:53.657409 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="northd" containerID="cri-o://8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e" gracePeriod=30 Mar 11 13:33:53 crc kubenswrapper[4923]: I0311 13:33:53.657457 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0" gracePeriod=30 Mar 11 13:33:53 crc kubenswrapper[4923]: I0311 13:33:53.657478 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="sbdb" containerID="cri-o://436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee" gracePeriod=30 Mar 11 13:33:53 crc kubenswrapper[4923]: I0311 13:33:53.657515 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="ovn-acl-logging" containerID="cri-o://a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d" gracePeriod=30 Mar 11 13:33:53 crc kubenswrapper[4923]: I0311 13:33:53.657503 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="kube-rbac-proxy-node" containerID="cri-o://eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e" gracePeriod=30 Mar 11 13:33:53 crc kubenswrapper[4923]: I0311 13:33:53.657371 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="nbdb" containerID="cri-o://21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34" gracePeriod=30 Mar 11 13:33:53 crc kubenswrapper[4923]: I0311 13:33:53.689685 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="ovnkube-controller" containerID="cri-o://6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f" gracePeriod=30 Mar 11 13:33:53 crc kubenswrapper[4923]: I0311 13:33:53.993444 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-552rb_df02f0a6-797a-4c04-a529-f0881241fd94/ovn-acl-logging/0.log" Mar 11 13:33:53 crc kubenswrapper[4923]: I0311 13:33:53.994550 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-552rb_df02f0a6-797a-4c04-a529-f0881241fd94/ovn-controller/0.log" Mar 11 13:33:53 crc kubenswrapper[4923]: I0311 13:33:53.995397 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.032653 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-var-lib-cni-networks-ovn-kubernetes\") pod \"df02f0a6-797a-4c04-a529-f0881241fd94\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.032947 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-run-systemd\") pod \"df02f0a6-797a-4c04-a529-f0881241fd94\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.033123 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-run-openvswitch\") pod \"df02f0a6-797a-4c04-a529-f0881241fd94\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.034549 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-var-lib-openvswitch\") pod \"df02f0a6-797a-4c04-a529-f0881241fd94\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.034642 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/df02f0a6-797a-4c04-a529-f0881241fd94-ovnkube-config\") pod \"df02f0a6-797a-4c04-a529-f0881241fd94\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.034694 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-node-log\") pod \"df02f0a6-797a-4c04-a529-f0881241fd94\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.034727 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-run-ovn-kubernetes\") pod \"df02f0a6-797a-4c04-a529-f0881241fd94\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.034771 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/df02f0a6-797a-4c04-a529-f0881241fd94-env-overrides\") pod \"df02f0a6-797a-4c04-a529-f0881241fd94\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.034807 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-systemd-units\") pod \"df02f0a6-797a-4c04-a529-f0881241fd94\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.034860 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-run-netns\") pod \"df02f0a6-797a-4c04-a529-f0881241fd94\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.034898 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-kubelet\") pod \"df02f0a6-797a-4c04-a529-f0881241fd94\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.034937 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-log-socket\") pod \"df02f0a6-797a-4c04-a529-f0881241fd94\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.034977 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/df02f0a6-797a-4c04-a529-f0881241fd94-ovn-node-metrics-cert\") pod \"df02f0a6-797a-4c04-a529-f0881241fd94\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.035010 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-etc-openvswitch\") pod \"df02f0a6-797a-4c04-a529-f0881241fd94\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.035050 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22njz\" (UniqueName: \"kubernetes.io/projected/df02f0a6-797a-4c04-a529-f0881241fd94-kube-api-access-22njz\") pod \"df02f0a6-797a-4c04-a529-f0881241fd94\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.035083 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/df02f0a6-797a-4c04-a529-f0881241fd94-ovnkube-script-lib\") pod \"df02f0a6-797a-4c04-a529-f0881241fd94\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.035111 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-slash\") pod \"df02f0a6-797a-4c04-a529-f0881241fd94\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.035147 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-run-ovn\") pod \"df02f0a6-797a-4c04-a529-f0881241fd94\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.035180 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-cni-netd\") pod \"df02f0a6-797a-4c04-a529-f0881241fd94\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.035212 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-cni-bin\") pod \"df02f0a6-797a-4c04-a529-f0881241fd94\" (UID: \"df02f0a6-797a-4c04-a529-f0881241fd94\") " Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.032973 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "df02f0a6-797a-4c04-a529-f0881241fd94" (UID: "df02f0a6-797a-4c04-a529-f0881241fd94"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.033384 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "df02f0a6-797a-4c04-a529-f0881241fd94" (UID: "df02f0a6-797a-4c04-a529-f0881241fd94"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.035908 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "df02f0a6-797a-4c04-a529-f0881241fd94" (UID: "df02f0a6-797a-4c04-a529-f0881241fd94"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.036020 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "df02f0a6-797a-4c04-a529-f0881241fd94" (UID: "df02f0a6-797a-4c04-a529-f0881241fd94"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.036948 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df02f0a6-797a-4c04-a529-f0881241fd94-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "df02f0a6-797a-4c04-a529-f0881241fd94" (UID: "df02f0a6-797a-4c04-a529-f0881241fd94"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.037030 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-node-log" (OuterVolumeSpecName: "node-log") pod "df02f0a6-797a-4c04-a529-f0881241fd94" (UID: "df02f0a6-797a-4c04-a529-f0881241fd94"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.037265 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "df02f0a6-797a-4c04-a529-f0881241fd94" (UID: "df02f0a6-797a-4c04-a529-f0881241fd94"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.040550 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-log-socket" (OuterVolumeSpecName: "log-socket") pod "df02f0a6-797a-4c04-a529-f0881241fd94" (UID: "df02f0a6-797a-4c04-a529-f0881241fd94"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.042141 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-slash" (OuterVolumeSpecName: "host-slash") pod "df02f0a6-797a-4c04-a529-f0881241fd94" (UID: "df02f0a6-797a-4c04-a529-f0881241fd94"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.042201 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "df02f0a6-797a-4c04-a529-f0881241fd94" (UID: "df02f0a6-797a-4c04-a529-f0881241fd94"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.042402 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "df02f0a6-797a-4c04-a529-f0881241fd94" (UID: "df02f0a6-797a-4c04-a529-f0881241fd94"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.042515 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "df02f0a6-797a-4c04-a529-f0881241fd94" (UID: "df02f0a6-797a-4c04-a529-f0881241fd94"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.042600 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "df02f0a6-797a-4c04-a529-f0881241fd94" (UID: "df02f0a6-797a-4c04-a529-f0881241fd94"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.042718 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "df02f0a6-797a-4c04-a529-f0881241fd94" (UID: "df02f0a6-797a-4c04-a529-f0881241fd94"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.043799 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df02f0a6-797a-4c04-a529-f0881241fd94-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "df02f0a6-797a-4c04-a529-f0881241fd94" (UID: "df02f0a6-797a-4c04-a529-f0881241fd94"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.044517 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "df02f0a6-797a-4c04-a529-f0881241fd94" (UID: "df02f0a6-797a-4c04-a529-f0881241fd94"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.046950 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df02f0a6-797a-4c04-a529-f0881241fd94-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "df02f0a6-797a-4c04-a529-f0881241fd94" (UID: "df02f0a6-797a-4c04-a529-f0881241fd94"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.047650 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df02f0a6-797a-4c04-a529-f0881241fd94-kube-api-access-22njz" (OuterVolumeSpecName: "kube-api-access-22njz") pod "df02f0a6-797a-4c04-a529-f0881241fd94" (UID: "df02f0a6-797a-4c04-a529-f0881241fd94"). InnerVolumeSpecName "kube-api-access-22njz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.048010 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df02f0a6-797a-4c04-a529-f0881241fd94-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "df02f0a6-797a-4c04-a529-f0881241fd94" (UID: "df02f0a6-797a-4c04-a529-f0881241fd94"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.060270 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "df02f0a6-797a-4c04-a529-f0881241fd94" (UID: "df02f0a6-797a-4c04-a529-f0881241fd94"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.067757 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7sww6"] Mar 11 13:33:54 crc kubenswrapper[4923]: E0311 13:33:54.068443 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="nbdb" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.068466 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="nbdb" Mar 11 13:33:54 crc kubenswrapper[4923]: E0311 13:33:54.068480 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="ovnkube-controller" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.068492 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="ovnkube-controller" Mar 11 13:33:54 crc kubenswrapper[4923]: E0311 13:33:54.068529 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="kube-rbac-proxy-ovn-metrics" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.068539 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="kube-rbac-proxy-ovn-metrics" Mar 11 13:33:54 crc kubenswrapper[4923]: E0311 13:33:54.068550 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="ovn-acl-logging" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.068558 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="ovn-acl-logging" Mar 11 13:33:54 crc kubenswrapper[4923]: E0311 13:33:54.068571 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="northd" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.068579 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="northd" Mar 11 13:33:54 crc kubenswrapper[4923]: E0311 13:33:54.068609 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39d4c595-2048-4b75-9c9e-e7d7d817afd7" containerName="oc" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.068618 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="39d4c595-2048-4b75-9c9e-e7d7d817afd7" containerName="oc" Mar 11 13:33:54 crc kubenswrapper[4923]: E0311 13:33:54.068634 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="kube-rbac-proxy-node" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.068642 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="kube-rbac-proxy-node" Mar 11 13:33:54 crc kubenswrapper[4923]: E0311 13:33:54.068655 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="sbdb" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.068662 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="sbdb" Mar 11 13:33:54 crc kubenswrapper[4923]: E0311 13:33:54.068698 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="kubecfg-setup" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.068707 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="kubecfg-setup" Mar 11 13:33:54 crc kubenswrapper[4923]: E0311 13:33:54.068719 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="ovn-controller" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.068727 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="ovn-controller" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.068878 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="ovn-acl-logging" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.068890 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="kube-rbac-proxy-node" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.068898 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="ovnkube-controller" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.068924 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="nbdb" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.068931 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="kube-rbac-proxy-ovn-metrics" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.068941 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="ovn-controller" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.068948 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="sbdb" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.068955 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" containerName="northd" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.068965 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="39d4c595-2048-4b75-9c9e-e7d7d817afd7" containerName="oc" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.072855 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.137280 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-host-cni-bin\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.137367 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-host-kubelet\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.137393 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-var-lib-openvswitch\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.137423 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-run-ovn\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.137470 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/389b2d5a-0c6a-4387-8e2b-379370c3eec7-ovnkube-config\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.137493 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/389b2d5a-0c6a-4387-8e2b-379370c3eec7-env-overrides\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.137537 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-host-run-ovn-kubernetes\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.137561 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-run-openvswitch\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.137615 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-etc-openvswitch\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.137643 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5zb6\" (UniqueName: \"kubernetes.io/projected/389b2d5a-0c6a-4387-8e2b-379370c3eec7-kube-api-access-h5zb6\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.137660 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-log-socket\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.137679 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.137704 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-run-systemd\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.137718 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-node-log\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.137732 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-systemd-units\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.137748 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-host-cni-netd\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.137828 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/389b2d5a-0c6a-4387-8e2b-379370c3eec7-ovnkube-script-lib\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.137941 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-host-slash\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.137974 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/389b2d5a-0c6a-4387-8e2b-379370c3eec7-ovn-node-metrics-cert\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.138011 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-host-run-netns\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.138138 4923 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/df02f0a6-797a-4c04-a529-f0881241fd94-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.138163 4923 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.138187 4923 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.138203 4923 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.138222 4923 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-log-socket\") on node \"crc\" DevicePath \"\"" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.138237 4923 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/df02f0a6-797a-4c04-a529-f0881241fd94-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.138256 4923 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.138270 4923 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/df02f0a6-797a-4c04-a529-f0881241fd94-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.138285 4923 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-slash\") on node \"crc\" DevicePath \"\"" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.138300 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22njz\" (UniqueName: \"kubernetes.io/projected/df02f0a6-797a-4c04-a529-f0881241fd94-kube-api-access-22njz\") on node \"crc\" DevicePath \"\"" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.138315 4923 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.138331 4923 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.138381 4923 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.138399 4923 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.138415 4923 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.138432 4923 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.138448 4923 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.138464 4923 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/df02f0a6-797a-4c04-a529-f0881241fd94-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.138481 4923 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-node-log\") on node \"crc\" DevicePath \"\"" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.138499 4923 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/df02f0a6-797a-4c04-a529-f0881241fd94-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.169561 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-552rb_df02f0a6-797a-4c04-a529-f0881241fd94/ovn-acl-logging/0.log" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170197 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-552rb_df02f0a6-797a-4c04-a529-f0881241fd94/ovn-controller/0.log" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170649 4923 generic.go:334] "Generic (PLEG): container finished" podID="df02f0a6-797a-4c04-a529-f0881241fd94" containerID="6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f" exitCode=0 Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170683 4923 generic.go:334] "Generic (PLEG): container finished" podID="df02f0a6-797a-4c04-a529-f0881241fd94" containerID="436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee" exitCode=0 Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170692 4923 generic.go:334] "Generic (PLEG): container finished" podID="df02f0a6-797a-4c04-a529-f0881241fd94" containerID="21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34" exitCode=0 Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170704 4923 generic.go:334] "Generic (PLEG): container finished" podID="df02f0a6-797a-4c04-a529-f0881241fd94" containerID="8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e" exitCode=0 Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170714 4923 generic.go:334] "Generic (PLEG): container finished" podID="df02f0a6-797a-4c04-a529-f0881241fd94" containerID="8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0" exitCode=0 Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170724 4923 generic.go:334] "Generic (PLEG): container finished" podID="df02f0a6-797a-4c04-a529-f0881241fd94" containerID="eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e" exitCode=0 Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170735 4923 generic.go:334] "Generic (PLEG): container finished" podID="df02f0a6-797a-4c04-a529-f0881241fd94" containerID="a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d" exitCode=143 Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170745 4923 generic.go:334] "Generic (PLEG): container finished" podID="df02f0a6-797a-4c04-a529-f0881241fd94" containerID="6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595" exitCode=143 Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170754 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" event={"ID":"df02f0a6-797a-4c04-a529-f0881241fd94","Type":"ContainerDied","Data":"6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170771 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170791 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" event={"ID":"df02f0a6-797a-4c04-a529-f0881241fd94","Type":"ContainerDied","Data":"436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170819 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" event={"ID":"df02f0a6-797a-4c04-a529-f0881241fd94","Type":"ContainerDied","Data":"21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170828 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" event={"ID":"df02f0a6-797a-4c04-a529-f0881241fd94","Type":"ContainerDied","Data":"8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170838 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" event={"ID":"df02f0a6-797a-4c04-a529-f0881241fd94","Type":"ContainerDied","Data":"8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170847 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" event={"ID":"df02f0a6-797a-4c04-a529-f0881241fd94","Type":"ContainerDied","Data":"eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170860 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170869 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170875 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170899 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" event={"ID":"df02f0a6-797a-4c04-a529-f0881241fd94","Type":"ContainerDied","Data":"a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170909 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170901 4923 scope.go:117] "RemoveContainer" containerID="6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.170917 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171063 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171081 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171091 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171101 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171110 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171119 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171128 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171150 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" event={"ID":"df02f0a6-797a-4c04-a529-f0881241fd94","Type":"ContainerDied","Data":"6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171175 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171184 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171192 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171199 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171206 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171214 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171220 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171227 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171235 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171245 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-552rb" event={"ID":"df02f0a6-797a-4c04-a529-f0881241fd94","Type":"ContainerDied","Data":"417cfa13eee32fab3ba6faf38c6585235ab1ae14743d4aa89ac68014dedfb0aa"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171255 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171264 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171271 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171278 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171286 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171293 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171301 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171308 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.171315 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.175655 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qfw4x_534d6af5-be40-40a6-accb-8b8b11cbf774/kube-multus/0.log" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.175690 4923 generic.go:334] "Generic (PLEG): container finished" podID="534d6af5-be40-40a6-accb-8b8b11cbf774" containerID="85cc2cd095a7a54377400f5d541dc109c15e09e067012495defd5aa6e2ec9ee3" exitCode=2 Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.175710 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qfw4x" event={"ID":"534d6af5-be40-40a6-accb-8b8b11cbf774","Type":"ContainerDied","Data":"85cc2cd095a7a54377400f5d541dc109c15e09e067012495defd5aa6e2ec9ee3"} Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.175955 4923 scope.go:117] "RemoveContainer" containerID="85cc2cd095a7a54377400f5d541dc109c15e09e067012495defd5aa6e2ec9ee3" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.205725 4923 scope.go:117] "RemoveContainer" containerID="436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.229385 4923 scope.go:117] "RemoveContainer" containerID="21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.239238 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/389b2d5a-0c6a-4387-8e2b-379370c3eec7-ovnkube-config\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.239271 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/389b2d5a-0c6a-4387-8e2b-379370c3eec7-env-overrides\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.239311 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-host-run-ovn-kubernetes\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.239328 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-run-openvswitch\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.239363 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-etc-openvswitch\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.239379 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5zb6\" (UniqueName: \"kubernetes.io/projected/389b2d5a-0c6a-4387-8e2b-379370c3eec7-kube-api-access-h5zb6\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.239395 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-log-socket\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.239418 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.239440 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-run-systemd\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.239454 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-node-log\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.239470 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-systemd-units\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.239487 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-host-cni-netd\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.239502 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/389b2d5a-0c6a-4387-8e2b-379370c3eec7-ovnkube-script-lib\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.239525 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-host-slash\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.239541 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/389b2d5a-0c6a-4387-8e2b-379370c3eec7-ovn-node-metrics-cert\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.239559 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-host-run-netns\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.239578 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-host-cni-bin\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.239598 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-host-kubelet\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.239613 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-var-lib-openvswitch\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.239630 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-run-ovn\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.240528 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-552rb"] Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.240993 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/389b2d5a-0c6a-4387-8e2b-379370c3eec7-ovnkube-config\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.241386 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/389b2d5a-0c6a-4387-8e2b-379370c3eec7-env-overrides\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.241649 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-host-run-ovn-kubernetes\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.241684 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-run-openvswitch\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.241842 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-host-cni-netd\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.241876 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-host-cni-bin\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.241902 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-host-run-netns\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.241985 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-etc-openvswitch\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.242049 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-run-ovn\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.242055 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.242110 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-var-lib-openvswitch\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.242092 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-host-kubelet\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.242156 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-log-socket\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.242189 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-run-systemd\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.242198 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/389b2d5a-0c6a-4387-8e2b-379370c3eec7-ovnkube-script-lib\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.242159 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-host-slash\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.242235 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-node-log\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.242237 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/389b2d5a-0c6a-4387-8e2b-379370c3eec7-systemd-units\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.249464 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/389b2d5a-0c6a-4387-8e2b-379370c3eec7-ovn-node-metrics-cert\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.254630 4923 scope.go:117] "RemoveContainer" containerID="8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.263474 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5zb6\" (UniqueName: \"kubernetes.io/projected/389b2d5a-0c6a-4387-8e2b-379370c3eec7-kube-api-access-h5zb6\") pod \"ovnkube-node-7sww6\" (UID: \"389b2d5a-0c6a-4387-8e2b-379370c3eec7\") " pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.267225 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-552rb"] Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.282833 4923 scope.go:117] "RemoveContainer" containerID="8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.301521 4923 scope.go:117] "RemoveContainer" containerID="eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.318558 4923 scope.go:117] "RemoveContainer" containerID="a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.338701 4923 scope.go:117] "RemoveContainer" containerID="6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.358282 4923 scope.go:117] "RemoveContainer" containerID="c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.384548 4923 scope.go:117] "RemoveContainer" containerID="6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f" Mar 11 13:33:54 crc kubenswrapper[4923]: E0311 13:33:54.385695 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f\": container with ID starting with 6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f not found: ID does not exist" containerID="6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.385891 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f"} err="failed to get container status \"6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f\": rpc error: code = NotFound desc = could not find container \"6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f\": container with ID starting with 6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.386007 4923 scope.go:117] "RemoveContainer" containerID="436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee" Mar 11 13:33:54 crc kubenswrapper[4923]: E0311 13:33:54.386748 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee\": container with ID starting with 436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee not found: ID does not exist" containerID="436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.386798 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee"} err="failed to get container status \"436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee\": rpc error: code = NotFound desc = could not find container \"436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee\": container with ID starting with 436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.386831 4923 scope.go:117] "RemoveContainer" containerID="21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34" Mar 11 13:33:54 crc kubenswrapper[4923]: E0311 13:33:54.387306 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34\": container with ID starting with 21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34 not found: ID does not exist" containerID="21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.387437 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34"} err="failed to get container status \"21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34\": rpc error: code = NotFound desc = could not find container \"21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34\": container with ID starting with 21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34 not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.387491 4923 scope.go:117] "RemoveContainer" containerID="8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e" Mar 11 13:33:54 crc kubenswrapper[4923]: E0311 13:33:54.388153 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e\": container with ID starting with 8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e not found: ID does not exist" containerID="8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.388191 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e"} err="failed to get container status \"8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e\": rpc error: code = NotFound desc = could not find container \"8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e\": container with ID starting with 8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.388213 4923 scope.go:117] "RemoveContainer" containerID="8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0" Mar 11 13:33:54 crc kubenswrapper[4923]: E0311 13:33:54.388600 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0\": container with ID starting with 8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0 not found: ID does not exist" containerID="8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.388720 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0"} err="failed to get container status \"8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0\": rpc error: code = NotFound desc = could not find container \"8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0\": container with ID starting with 8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0 not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.388814 4923 scope.go:117] "RemoveContainer" containerID="eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e" Mar 11 13:33:54 crc kubenswrapper[4923]: E0311 13:33:54.390465 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e\": container with ID starting with eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e not found: ID does not exist" containerID="eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.390495 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e"} err="failed to get container status \"eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e\": rpc error: code = NotFound desc = could not find container \"eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e\": container with ID starting with eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.390514 4923 scope.go:117] "RemoveContainer" containerID="a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d" Mar 11 13:33:54 crc kubenswrapper[4923]: E0311 13:33:54.390882 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d\": container with ID starting with a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d not found: ID does not exist" containerID="a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.390994 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d"} err="failed to get container status \"a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d\": rpc error: code = NotFound desc = could not find container \"a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d\": container with ID starting with a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.391087 4923 scope.go:117] "RemoveContainer" containerID="6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.391503 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:33:54 crc kubenswrapper[4923]: E0311 13:33:54.391588 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595\": container with ID starting with 6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595 not found: ID does not exist" containerID="6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.391625 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595"} err="failed to get container status \"6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595\": rpc error: code = NotFound desc = could not find container \"6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595\": container with ID starting with 6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595 not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.391652 4923 scope.go:117] "RemoveContainer" containerID="c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739" Mar 11 13:33:54 crc kubenswrapper[4923]: E0311 13:33:54.392485 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739\": container with ID starting with c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739 not found: ID does not exist" containerID="c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.392574 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739"} err="failed to get container status \"c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739\": rpc error: code = NotFound desc = could not find container \"c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739\": container with ID starting with c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739 not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.392632 4923 scope.go:117] "RemoveContainer" containerID="6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.393132 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f"} err="failed to get container status \"6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f\": rpc error: code = NotFound desc = could not find container \"6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f\": container with ID starting with 6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.393190 4923 scope.go:117] "RemoveContainer" containerID="436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.393697 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee"} err="failed to get container status \"436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee\": rpc error: code = NotFound desc = could not find container \"436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee\": container with ID starting with 436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.393753 4923 scope.go:117] "RemoveContainer" containerID="21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.394138 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34"} err="failed to get container status \"21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34\": rpc error: code = NotFound desc = could not find container \"21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34\": container with ID starting with 21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34 not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.394187 4923 scope.go:117] "RemoveContainer" containerID="8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.394521 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e"} err="failed to get container status \"8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e\": rpc error: code = NotFound desc = could not find container \"8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e\": container with ID starting with 8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.394561 4923 scope.go:117] "RemoveContainer" containerID="8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.394913 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0"} err="failed to get container status \"8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0\": rpc error: code = NotFound desc = could not find container \"8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0\": container with ID starting with 8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0 not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.394977 4923 scope.go:117] "RemoveContainer" containerID="eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.395553 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e"} err="failed to get container status \"eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e\": rpc error: code = NotFound desc = could not find container \"eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e\": container with ID starting with eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.395607 4923 scope.go:117] "RemoveContainer" containerID="a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.396069 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d"} err="failed to get container status \"a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d\": rpc error: code = NotFound desc = could not find container \"a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d\": container with ID starting with a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.396119 4923 scope.go:117] "RemoveContainer" containerID="6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.396511 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595"} err="failed to get container status \"6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595\": rpc error: code = NotFound desc = could not find container \"6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595\": container with ID starting with 6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595 not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.396551 4923 scope.go:117] "RemoveContainer" containerID="c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.396888 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739"} err="failed to get container status \"c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739\": rpc error: code = NotFound desc = could not find container \"c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739\": container with ID starting with c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739 not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.396945 4923 scope.go:117] "RemoveContainer" containerID="6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.397394 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f"} err="failed to get container status \"6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f\": rpc error: code = NotFound desc = could not find container \"6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f\": container with ID starting with 6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.397444 4923 scope.go:117] "RemoveContainer" containerID="436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.397930 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee"} err="failed to get container status \"436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee\": rpc error: code = NotFound desc = could not find container \"436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee\": container with ID starting with 436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.398043 4923 scope.go:117] "RemoveContainer" containerID="21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.398639 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34"} err="failed to get container status \"21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34\": rpc error: code = NotFound desc = could not find container \"21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34\": container with ID starting with 21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34 not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.398691 4923 scope.go:117] "RemoveContainer" containerID="8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.399312 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e"} err="failed to get container status \"8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e\": rpc error: code = NotFound desc = could not find container \"8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e\": container with ID starting with 8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.399415 4923 scope.go:117] "RemoveContainer" containerID="8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.399797 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0"} err="failed to get container status \"8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0\": rpc error: code = NotFound desc = could not find container \"8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0\": container with ID starting with 8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0 not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.399855 4923 scope.go:117] "RemoveContainer" containerID="eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.400290 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e"} err="failed to get container status \"eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e\": rpc error: code = NotFound desc = could not find container \"eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e\": container with ID starting with eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.400424 4923 scope.go:117] "RemoveContainer" containerID="a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.400756 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d"} err="failed to get container status \"a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d\": rpc error: code = NotFound desc = could not find container \"a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d\": container with ID starting with a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.400864 4923 scope.go:117] "RemoveContainer" containerID="6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.401307 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595"} err="failed to get container status \"6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595\": rpc error: code = NotFound desc = could not find container \"6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595\": container with ID starting with 6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595 not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.401455 4923 scope.go:117] "RemoveContainer" containerID="c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.401795 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739"} err="failed to get container status \"c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739\": rpc error: code = NotFound desc = could not find container \"c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739\": container with ID starting with c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739 not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.401907 4923 scope.go:117] "RemoveContainer" containerID="6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.402301 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f"} err="failed to get container status \"6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f\": rpc error: code = NotFound desc = could not find container \"6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f\": container with ID starting with 6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.402402 4923 scope.go:117] "RemoveContainer" containerID="436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.402772 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee"} err="failed to get container status \"436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee\": rpc error: code = NotFound desc = could not find container \"436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee\": container with ID starting with 436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.402872 4923 scope.go:117] "RemoveContainer" containerID="21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.403409 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34"} err="failed to get container status \"21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34\": rpc error: code = NotFound desc = could not find container \"21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34\": container with ID starting with 21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34 not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.403483 4923 scope.go:117] "RemoveContainer" containerID="8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.403824 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e"} err="failed to get container status \"8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e\": rpc error: code = NotFound desc = could not find container \"8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e\": container with ID starting with 8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.403881 4923 scope.go:117] "RemoveContainer" containerID="8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.404532 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0"} err="failed to get container status \"8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0\": rpc error: code = NotFound desc = could not find container \"8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0\": container with ID starting with 8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0 not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.404623 4923 scope.go:117] "RemoveContainer" containerID="eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.405190 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e"} err="failed to get container status \"eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e\": rpc error: code = NotFound desc = could not find container \"eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e\": container with ID starting with eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.405392 4923 scope.go:117] "RemoveContainer" containerID="a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.406294 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d"} err="failed to get container status \"a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d\": rpc error: code = NotFound desc = could not find container \"a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d\": container with ID starting with a5f273be41ff359f0a49a408d225306e1d5faa144e40e8b3b2ce03631efcb55d not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.406420 4923 scope.go:117] "RemoveContainer" containerID="6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.406860 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595"} err="failed to get container status \"6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595\": rpc error: code = NotFound desc = could not find container \"6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595\": container with ID starting with 6d5913737663fcaeabbf8c90d5167c1ac054b47bbf720bcf8cd9155b34522595 not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.406926 4923 scope.go:117] "RemoveContainer" containerID="c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.407233 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739"} err="failed to get container status \"c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739\": rpc error: code = NotFound desc = could not find container \"c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739\": container with ID starting with c6feb27c729bbebc8074e0951b6b1ac4181bd4472e50c9abe0fb72edc2115739 not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.407259 4923 scope.go:117] "RemoveContainer" containerID="6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.407692 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f"} err="failed to get container status \"6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f\": rpc error: code = NotFound desc = could not find container \"6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f\": container with ID starting with 6fcb038d3daaf38dd96ce66337b3e8f4a5744e43dc49d61fff23d927757da82f not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.407722 4923 scope.go:117] "RemoveContainer" containerID="436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.408107 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee"} err="failed to get container status \"436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee\": rpc error: code = NotFound desc = could not find container \"436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee\": container with ID starting with 436cc8e8a49b42155e677e587647e57db423c3745cafee6ae137981d11c33bee not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.408190 4923 scope.go:117] "RemoveContainer" containerID="21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.408608 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34"} err="failed to get container status \"21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34\": rpc error: code = NotFound desc = could not find container \"21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34\": container with ID starting with 21a5dfbff8018ff9e603d8f91eb7592801f2047d262001e8f75b2be65855cd34 not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.408642 4923 scope.go:117] "RemoveContainer" containerID="8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.409014 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e"} err="failed to get container status \"8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e\": rpc error: code = NotFound desc = could not find container \"8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e\": container with ID starting with 8197ec9e8f271cbcd5dd2ca66c8a11f76564ec2e42b70c79d7f102a617966d8e not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.409078 4923 scope.go:117] "RemoveContainer" containerID="8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.410520 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0"} err="failed to get container status \"8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0\": rpc error: code = NotFound desc = could not find container \"8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0\": container with ID starting with 8fb90dec842ed74665737954860b0869672f8fee5dfb1e4da3c394ac227065e0 not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.410560 4923 scope.go:117] "RemoveContainer" containerID="eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e" Mar 11 13:33:54 crc kubenswrapper[4923]: I0311 13:33:54.411212 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e"} err="failed to get container status \"eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e\": rpc error: code = NotFound desc = could not find container \"eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e\": container with ID starting with eb65b0a39dda11524c70e7a954bf7187b55e97baf727e41ab5e0bd5dafb98f0e not found: ID does not exist" Mar 11 13:33:54 crc kubenswrapper[4923]: W0311 13:33:54.419928 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod389b2d5a_0c6a_4387_8e2b_379370c3eec7.slice/crio-e465e7f2317758fb9357b3b1fbee03ca88acd2aeeb85d7782ea6708bd57fabed WatchSource:0}: Error finding container e465e7f2317758fb9357b3b1fbee03ca88acd2aeeb85d7782ea6708bd57fabed: Status 404 returned error can't find the container with id e465e7f2317758fb9357b3b1fbee03ca88acd2aeeb85d7782ea6708bd57fabed Mar 11 13:33:55 crc kubenswrapper[4923]: I0311 13:33:55.034160 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df02f0a6-797a-4c04-a529-f0881241fd94" path="/var/lib/kubelet/pods/df02f0a6-797a-4c04-a529-f0881241fd94/volumes" Mar 11 13:33:55 crc kubenswrapper[4923]: I0311 13:33:55.185289 4923 generic.go:334] "Generic (PLEG): container finished" podID="389b2d5a-0c6a-4387-8e2b-379370c3eec7" containerID="662e7784b8dfe44d2d7d28236043cf7c287ee4c5a370e4bc9e20db19538e5608" exitCode=0 Mar 11 13:33:55 crc kubenswrapper[4923]: I0311 13:33:55.185404 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" event={"ID":"389b2d5a-0c6a-4387-8e2b-379370c3eec7","Type":"ContainerDied","Data":"662e7784b8dfe44d2d7d28236043cf7c287ee4c5a370e4bc9e20db19538e5608"} Mar 11 13:33:55 crc kubenswrapper[4923]: I0311 13:33:55.185488 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" event={"ID":"389b2d5a-0c6a-4387-8e2b-379370c3eec7","Type":"ContainerStarted","Data":"e465e7f2317758fb9357b3b1fbee03ca88acd2aeeb85d7782ea6708bd57fabed"} Mar 11 13:33:55 crc kubenswrapper[4923]: I0311 13:33:55.190024 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qfw4x_534d6af5-be40-40a6-accb-8b8b11cbf774/kube-multus/0.log" Mar 11 13:33:55 crc kubenswrapper[4923]: I0311 13:33:55.190184 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qfw4x" event={"ID":"534d6af5-be40-40a6-accb-8b8b11cbf774","Type":"ContainerStarted","Data":"4c6301c64060bd05dba9342d368d9e7a171da9c5ebb79d2ffb59682509351cb5"} Mar 11 13:33:56 crc kubenswrapper[4923]: I0311 13:33:56.212770 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" event={"ID":"389b2d5a-0c6a-4387-8e2b-379370c3eec7","Type":"ContainerStarted","Data":"c6ecc6d66091c6877b3c42320be3f61ee2d7eabafcbd550f6e5ff3df510697a1"} Mar 11 13:33:56 crc kubenswrapper[4923]: I0311 13:33:56.213169 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" event={"ID":"389b2d5a-0c6a-4387-8e2b-379370c3eec7","Type":"ContainerStarted","Data":"ea68f7ddec9b8720b17d5ac6591920ca877b637e1636c3c78ef2b108d69f8090"} Mar 11 13:33:56 crc kubenswrapper[4923]: I0311 13:33:56.213190 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" event={"ID":"389b2d5a-0c6a-4387-8e2b-379370c3eec7","Type":"ContainerStarted","Data":"b91f967996c93d08fa927b42c669ad13b1ad9757af9fce983537b07864b2532e"} Mar 11 13:33:56 crc kubenswrapper[4923]: I0311 13:33:56.213209 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" event={"ID":"389b2d5a-0c6a-4387-8e2b-379370c3eec7","Type":"ContainerStarted","Data":"937d09a86191aa7e87d1d31e58e99436ebfa999f43eb5375a08b9670f0007c6f"} Mar 11 13:33:56 crc kubenswrapper[4923]: I0311 13:33:56.213226 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" event={"ID":"389b2d5a-0c6a-4387-8e2b-379370c3eec7","Type":"ContainerStarted","Data":"0845951ca07037a9f2f8787d202930ff4a9a48cb44b4e95e4f28fa01309e7689"} Mar 11 13:33:57 crc kubenswrapper[4923]: I0311 13:33:57.224981 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" event={"ID":"389b2d5a-0c6a-4387-8e2b-379370c3eec7","Type":"ContainerStarted","Data":"c2d2e8cdf121847a00bc08c36e40b22333cde318f3f2b51446bf7385ef376ea3"} Mar 11 13:33:59 crc kubenswrapper[4923]: I0311 13:33:59.243387 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" event={"ID":"389b2d5a-0c6a-4387-8e2b-379370c3eec7","Type":"ContainerStarted","Data":"4ad7745d4164f01b79a26bad0fc4ff784bec76673bc203270b28c4f6dfdb4145"} Mar 11 13:34:00 crc kubenswrapper[4923]: I0311 13:34:00.142337 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29553934-p2kch"] Mar 11 13:34:00 crc kubenswrapper[4923]: I0311 13:34:00.143091 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553934-p2kch" Mar 11 13:34:00 crc kubenswrapper[4923]: I0311 13:34:00.145869 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 11 13:34:00 crc kubenswrapper[4923]: I0311 13:34:00.146086 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-nxck7" Mar 11 13:34:00 crc kubenswrapper[4923]: I0311 13:34:00.147955 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 11 13:34:00 crc kubenswrapper[4923]: I0311 13:34:00.316588 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsfdc\" (UniqueName: \"kubernetes.io/projected/ecd0e139-902f-425f-ba54-08eac773e2ae-kube-api-access-vsfdc\") pod \"auto-csr-approver-29553934-p2kch\" (UID: \"ecd0e139-902f-425f-ba54-08eac773e2ae\") " pod="openshift-infra/auto-csr-approver-29553934-p2kch" Mar 11 13:34:00 crc kubenswrapper[4923]: I0311 13:34:00.417524 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsfdc\" (UniqueName: \"kubernetes.io/projected/ecd0e139-902f-425f-ba54-08eac773e2ae-kube-api-access-vsfdc\") pod \"auto-csr-approver-29553934-p2kch\" (UID: \"ecd0e139-902f-425f-ba54-08eac773e2ae\") " pod="openshift-infra/auto-csr-approver-29553934-p2kch" Mar 11 13:34:00 crc kubenswrapper[4923]: I0311 13:34:00.444217 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsfdc\" (UniqueName: \"kubernetes.io/projected/ecd0e139-902f-425f-ba54-08eac773e2ae-kube-api-access-vsfdc\") pod \"auto-csr-approver-29553934-p2kch\" (UID: \"ecd0e139-902f-425f-ba54-08eac773e2ae\") " pod="openshift-infra/auto-csr-approver-29553934-p2kch" Mar 11 13:34:00 crc kubenswrapper[4923]: I0311 13:34:00.467793 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553934-p2kch" Mar 11 13:34:00 crc kubenswrapper[4923]: E0311 13:34:00.487253 4923 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29553934-p2kch_openshift-infra_ecd0e139-902f-425f-ba54-08eac773e2ae_0(1b0761fe9aabb5b0d62a9b2509f4942a5c5c909c274d338b2764af4f06764505): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 11 13:34:00 crc kubenswrapper[4923]: E0311 13:34:00.487327 4923 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29553934-p2kch_openshift-infra_ecd0e139-902f-425f-ba54-08eac773e2ae_0(1b0761fe9aabb5b0d62a9b2509f4942a5c5c909c274d338b2764af4f06764505): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29553934-p2kch" Mar 11 13:34:00 crc kubenswrapper[4923]: E0311 13:34:00.487367 4923 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29553934-p2kch_openshift-infra_ecd0e139-902f-425f-ba54-08eac773e2ae_0(1b0761fe9aabb5b0d62a9b2509f4942a5c5c909c274d338b2764af4f06764505): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29553934-p2kch" Mar 11 13:34:00 crc kubenswrapper[4923]: E0311 13:34:00.487422 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"auto-csr-approver-29553934-p2kch_openshift-infra(ecd0e139-902f-425f-ba54-08eac773e2ae)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"auto-csr-approver-29553934-p2kch_openshift-infra(ecd0e139-902f-425f-ba54-08eac773e2ae)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29553934-p2kch_openshift-infra_ecd0e139-902f-425f-ba54-08eac773e2ae_0(1b0761fe9aabb5b0d62a9b2509f4942a5c5c909c274d338b2764af4f06764505): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-infra/auto-csr-approver-29553934-p2kch" podUID="ecd0e139-902f-425f-ba54-08eac773e2ae" Mar 11 13:34:01 crc kubenswrapper[4923]: I0311 13:34:01.262383 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" event={"ID":"389b2d5a-0c6a-4387-8e2b-379370c3eec7","Type":"ContainerStarted","Data":"8e5995063f042f38007f06a458b7da673f0d4e3bce9982997ef80296abcc0d05"} Mar 11 13:34:01 crc kubenswrapper[4923]: I0311 13:34:01.262849 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:34:01 crc kubenswrapper[4923]: I0311 13:34:01.262874 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:34:01 crc kubenswrapper[4923]: I0311 13:34:01.262891 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:34:01 crc kubenswrapper[4923]: I0311 13:34:01.306631 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:34:01 crc kubenswrapper[4923]: I0311 13:34:01.307765 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" podStartSLOduration=7.30774063 podStartE2EDuration="7.30774063s" podCreationTimestamp="2026-03-11 13:33:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:34:01.301814619 +0000 UTC m=+688.983716363" watchObservedRunningTime="2026-03-11 13:34:01.30774063 +0000 UTC m=+688.989642354" Mar 11 13:34:01 crc kubenswrapper[4923]: I0311 13:34:01.307858 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:34:01 crc kubenswrapper[4923]: I0311 13:34:01.717317 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553934-p2kch"] Mar 11 13:34:01 crc kubenswrapper[4923]: I0311 13:34:01.717623 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553934-p2kch" Mar 11 13:34:01 crc kubenswrapper[4923]: I0311 13:34:01.718162 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553934-p2kch" Mar 11 13:34:01 crc kubenswrapper[4923]: E0311 13:34:01.771432 4923 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29553934-p2kch_openshift-infra_ecd0e139-902f-425f-ba54-08eac773e2ae_0(692b7ce6b2b29f466706a02f5cd30baad978dafdb6e201d06901940f79f28a33): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 11 13:34:01 crc kubenswrapper[4923]: E0311 13:34:01.771506 4923 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29553934-p2kch_openshift-infra_ecd0e139-902f-425f-ba54-08eac773e2ae_0(692b7ce6b2b29f466706a02f5cd30baad978dafdb6e201d06901940f79f28a33): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29553934-p2kch" Mar 11 13:34:01 crc kubenswrapper[4923]: E0311 13:34:01.771553 4923 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29553934-p2kch_openshift-infra_ecd0e139-902f-425f-ba54-08eac773e2ae_0(692b7ce6b2b29f466706a02f5cd30baad978dafdb6e201d06901940f79f28a33): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29553934-p2kch" Mar 11 13:34:01 crc kubenswrapper[4923]: E0311 13:34:01.771612 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"auto-csr-approver-29553934-p2kch_openshift-infra(ecd0e139-902f-425f-ba54-08eac773e2ae)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"auto-csr-approver-29553934-p2kch_openshift-infra(ecd0e139-902f-425f-ba54-08eac773e2ae)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29553934-p2kch_openshift-infra_ecd0e139-902f-425f-ba54-08eac773e2ae_0(692b7ce6b2b29f466706a02f5cd30baad978dafdb6e201d06901940f79f28a33): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-infra/auto-csr-approver-29553934-p2kch" podUID="ecd0e139-902f-425f-ba54-08eac773e2ae" Mar 11 13:34:11 crc kubenswrapper[4923]: I0311 13:34:11.932566 4923 patch_prober.go:28] interesting pod/machine-config-daemon-fsfnw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 11 13:34:11 crc kubenswrapper[4923]: I0311 13:34:11.933539 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 11 13:34:14 crc kubenswrapper[4923]: I0311 13:34:14.021474 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553934-p2kch" Mar 11 13:34:14 crc kubenswrapper[4923]: I0311 13:34:14.022672 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553934-p2kch" Mar 11 13:34:14 crc kubenswrapper[4923]: I0311 13:34:14.272831 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553934-p2kch"] Mar 11 13:34:14 crc kubenswrapper[4923]: W0311 13:34:14.280251 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecd0e139_902f_425f_ba54_08eac773e2ae.slice/crio-533a8b92e6443481a444fdb7fcc445abe57ac82ad4ef09d2f5df6e3a5e538b4f WatchSource:0}: Error finding container 533a8b92e6443481a444fdb7fcc445abe57ac82ad4ef09d2f5df6e3a5e538b4f: Status 404 returned error can't find the container with id 533a8b92e6443481a444fdb7fcc445abe57ac82ad4ef09d2f5df6e3a5e538b4f Mar 11 13:34:14 crc kubenswrapper[4923]: I0311 13:34:14.353036 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553934-p2kch" event={"ID":"ecd0e139-902f-425f-ba54-08eac773e2ae","Type":"ContainerStarted","Data":"533a8b92e6443481a444fdb7fcc445abe57ac82ad4ef09d2f5df6e3a5e538b4f"} Mar 11 13:34:16 crc kubenswrapper[4923]: I0311 13:34:16.371568 4923 generic.go:334] "Generic (PLEG): container finished" podID="ecd0e139-902f-425f-ba54-08eac773e2ae" containerID="922fad3b003dde7f78c678afe06de492a98a65c1a6db57c64cf032ab8918dc52" exitCode=0 Mar 11 13:34:16 crc kubenswrapper[4923]: I0311 13:34:16.371725 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553934-p2kch" event={"ID":"ecd0e139-902f-425f-ba54-08eac773e2ae","Type":"ContainerDied","Data":"922fad3b003dde7f78c678afe06de492a98a65c1a6db57c64cf032ab8918dc52"} Mar 11 13:34:17 crc kubenswrapper[4923]: I0311 13:34:17.738029 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553934-p2kch" Mar 11 13:34:17 crc kubenswrapper[4923]: I0311 13:34:17.869549 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsfdc\" (UniqueName: \"kubernetes.io/projected/ecd0e139-902f-425f-ba54-08eac773e2ae-kube-api-access-vsfdc\") pod \"ecd0e139-902f-425f-ba54-08eac773e2ae\" (UID: \"ecd0e139-902f-425f-ba54-08eac773e2ae\") " Mar 11 13:34:17 crc kubenswrapper[4923]: I0311 13:34:17.879809 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecd0e139-902f-425f-ba54-08eac773e2ae-kube-api-access-vsfdc" (OuterVolumeSpecName: "kube-api-access-vsfdc") pod "ecd0e139-902f-425f-ba54-08eac773e2ae" (UID: "ecd0e139-902f-425f-ba54-08eac773e2ae"). InnerVolumeSpecName "kube-api-access-vsfdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:34:17 crc kubenswrapper[4923]: I0311 13:34:17.971581 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsfdc\" (UniqueName: \"kubernetes.io/projected/ecd0e139-902f-425f-ba54-08eac773e2ae-kube-api-access-vsfdc\") on node \"crc\" DevicePath \"\"" Mar 11 13:34:18 crc kubenswrapper[4923]: I0311 13:34:18.391165 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553934-p2kch" event={"ID":"ecd0e139-902f-425f-ba54-08eac773e2ae","Type":"ContainerDied","Data":"533a8b92e6443481a444fdb7fcc445abe57ac82ad4ef09d2f5df6e3a5e538b4f"} Mar 11 13:34:18 crc kubenswrapper[4923]: I0311 13:34:18.391217 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="533a8b92e6443481a444fdb7fcc445abe57ac82ad4ef09d2f5df6e3a5e538b4f" Mar 11 13:34:18 crc kubenswrapper[4923]: I0311 13:34:18.391230 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553934-p2kch" Mar 11 13:34:18 crc kubenswrapper[4923]: I0311 13:34:18.847428 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29553928-db6gx"] Mar 11 13:34:18 crc kubenswrapper[4923]: I0311 13:34:18.859257 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29553928-db6gx"] Mar 11 13:34:19 crc kubenswrapper[4923]: I0311 13:34:19.028675 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d7c3568-b74c-4389-99e9-4bf2b62bff75" path="/var/lib/kubelet/pods/4d7c3568-b74c-4389-99e9-4bf2b62bff75/volumes" Mar 11 13:34:20 crc kubenswrapper[4923]: I0311 13:34:20.468920 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww"] Mar 11 13:34:20 crc kubenswrapper[4923]: E0311 13:34:20.469706 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecd0e139-902f-425f-ba54-08eac773e2ae" containerName="oc" Mar 11 13:34:20 crc kubenswrapper[4923]: I0311 13:34:20.469725 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecd0e139-902f-425f-ba54-08eac773e2ae" containerName="oc" Mar 11 13:34:20 crc kubenswrapper[4923]: I0311 13:34:20.469866 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecd0e139-902f-425f-ba54-08eac773e2ae" containerName="oc" Mar 11 13:34:20 crc kubenswrapper[4923]: I0311 13:34:20.471057 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww" Mar 11 13:34:20 crc kubenswrapper[4923]: I0311 13:34:20.476039 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 11 13:34:20 crc kubenswrapper[4923]: I0311 13:34:20.478541 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww"] Mar 11 13:34:20 crc kubenswrapper[4923]: I0311 13:34:20.602170 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/45210d52-d52e-4809-b571-ca076d6e9d84-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww\" (UID: \"45210d52-d52e-4809-b571-ca076d6e9d84\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww" Mar 11 13:34:20 crc kubenswrapper[4923]: I0311 13:34:20.602308 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/45210d52-d52e-4809-b571-ca076d6e9d84-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww\" (UID: \"45210d52-d52e-4809-b571-ca076d6e9d84\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww" Mar 11 13:34:20 crc kubenswrapper[4923]: I0311 13:34:20.602471 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v99kb\" (UniqueName: \"kubernetes.io/projected/45210d52-d52e-4809-b571-ca076d6e9d84-kube-api-access-v99kb\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww\" (UID: \"45210d52-d52e-4809-b571-ca076d6e9d84\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww" Mar 11 13:34:20 crc kubenswrapper[4923]: I0311 13:34:20.703493 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v99kb\" (UniqueName: \"kubernetes.io/projected/45210d52-d52e-4809-b571-ca076d6e9d84-kube-api-access-v99kb\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww\" (UID: \"45210d52-d52e-4809-b571-ca076d6e9d84\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww" Mar 11 13:34:20 crc kubenswrapper[4923]: I0311 13:34:20.703562 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/45210d52-d52e-4809-b571-ca076d6e9d84-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww\" (UID: \"45210d52-d52e-4809-b571-ca076d6e9d84\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww" Mar 11 13:34:20 crc kubenswrapper[4923]: I0311 13:34:20.703602 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/45210d52-d52e-4809-b571-ca076d6e9d84-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww\" (UID: \"45210d52-d52e-4809-b571-ca076d6e9d84\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww" Mar 11 13:34:20 crc kubenswrapper[4923]: I0311 13:34:20.704196 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/45210d52-d52e-4809-b571-ca076d6e9d84-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww\" (UID: \"45210d52-d52e-4809-b571-ca076d6e9d84\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww" Mar 11 13:34:20 crc kubenswrapper[4923]: I0311 13:34:20.704777 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/45210d52-d52e-4809-b571-ca076d6e9d84-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww\" (UID: \"45210d52-d52e-4809-b571-ca076d6e9d84\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww" Mar 11 13:34:20 crc kubenswrapper[4923]: I0311 13:34:20.731570 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v99kb\" (UniqueName: \"kubernetes.io/projected/45210d52-d52e-4809-b571-ca076d6e9d84-kube-api-access-v99kb\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww\" (UID: \"45210d52-d52e-4809-b571-ca076d6e9d84\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww" Mar 11 13:34:20 crc kubenswrapper[4923]: I0311 13:34:20.798303 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww" Mar 11 13:34:21 crc kubenswrapper[4923]: I0311 13:34:21.249231 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww"] Mar 11 13:34:21 crc kubenswrapper[4923]: I0311 13:34:21.411258 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww" event={"ID":"45210d52-d52e-4809-b571-ca076d6e9d84","Type":"ContainerStarted","Data":"4fb7a276a20f5a532fb894d545935ed5bf4662c1b858644c2afbcabd45531dc4"} Mar 11 13:34:21 crc kubenswrapper[4923]: I0311 13:34:21.411390 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww" event={"ID":"45210d52-d52e-4809-b571-ca076d6e9d84","Type":"ContainerStarted","Data":"7eb205dc406e5022702720ac1e2018644f6c4e421c743ab5e2c34991323d39b0"} Mar 11 13:34:22 crc kubenswrapper[4923]: I0311 13:34:22.423823 4923 generic.go:334] "Generic (PLEG): container finished" podID="45210d52-d52e-4809-b571-ca076d6e9d84" containerID="4fb7a276a20f5a532fb894d545935ed5bf4662c1b858644c2afbcabd45531dc4" exitCode=0 Mar 11 13:34:22 crc kubenswrapper[4923]: I0311 13:34:22.423974 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww" event={"ID":"45210d52-d52e-4809-b571-ca076d6e9d84","Type":"ContainerDied","Data":"4fb7a276a20f5a532fb894d545935ed5bf4662c1b858644c2afbcabd45531dc4"} Mar 11 13:34:24 crc kubenswrapper[4923]: I0311 13:34:24.429212 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7sww6" Mar 11 13:34:24 crc kubenswrapper[4923]: I0311 13:34:24.438469 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww" event={"ID":"45210d52-d52e-4809-b571-ca076d6e9d84","Type":"ContainerStarted","Data":"57056373d6793bec2fe8da5c831e499aad3738051450a0ad8c7bcd982979ba16"} Mar 11 13:34:25 crc kubenswrapper[4923]: I0311 13:34:25.448764 4923 generic.go:334] "Generic (PLEG): container finished" podID="45210d52-d52e-4809-b571-ca076d6e9d84" containerID="57056373d6793bec2fe8da5c831e499aad3738051450a0ad8c7bcd982979ba16" exitCode=0 Mar 11 13:34:25 crc kubenswrapper[4923]: I0311 13:34:25.448833 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww" event={"ID":"45210d52-d52e-4809-b571-ca076d6e9d84","Type":"ContainerDied","Data":"57056373d6793bec2fe8da5c831e499aad3738051450a0ad8c7bcd982979ba16"} Mar 11 13:34:26 crc kubenswrapper[4923]: I0311 13:34:26.459459 4923 generic.go:334] "Generic (PLEG): container finished" podID="45210d52-d52e-4809-b571-ca076d6e9d84" containerID="bb9fd11292d57f8b54e61880ddd30df8d4e303203f9c68838c18f7f93fff47a1" exitCode=0 Mar 11 13:34:26 crc kubenswrapper[4923]: I0311 13:34:26.459525 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww" event={"ID":"45210d52-d52e-4809-b571-ca076d6e9d84","Type":"ContainerDied","Data":"bb9fd11292d57f8b54e61880ddd30df8d4e303203f9c68838c18f7f93fff47a1"} Mar 11 13:34:27 crc kubenswrapper[4923]: I0311 13:34:27.785632 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww" Mar 11 13:34:27 crc kubenswrapper[4923]: I0311 13:34:27.897117 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v99kb\" (UniqueName: \"kubernetes.io/projected/45210d52-d52e-4809-b571-ca076d6e9d84-kube-api-access-v99kb\") pod \"45210d52-d52e-4809-b571-ca076d6e9d84\" (UID: \"45210d52-d52e-4809-b571-ca076d6e9d84\") " Mar 11 13:34:27 crc kubenswrapper[4923]: I0311 13:34:27.897307 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/45210d52-d52e-4809-b571-ca076d6e9d84-util\") pod \"45210d52-d52e-4809-b571-ca076d6e9d84\" (UID: \"45210d52-d52e-4809-b571-ca076d6e9d84\") " Mar 11 13:34:27 crc kubenswrapper[4923]: I0311 13:34:27.897640 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/45210d52-d52e-4809-b571-ca076d6e9d84-bundle\") pod \"45210d52-d52e-4809-b571-ca076d6e9d84\" (UID: \"45210d52-d52e-4809-b571-ca076d6e9d84\") " Mar 11 13:34:27 crc kubenswrapper[4923]: I0311 13:34:27.899694 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45210d52-d52e-4809-b571-ca076d6e9d84-bundle" (OuterVolumeSpecName: "bundle") pod "45210d52-d52e-4809-b571-ca076d6e9d84" (UID: "45210d52-d52e-4809-b571-ca076d6e9d84"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:34:27 crc kubenswrapper[4923]: I0311 13:34:27.906269 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45210d52-d52e-4809-b571-ca076d6e9d84-kube-api-access-v99kb" (OuterVolumeSpecName: "kube-api-access-v99kb") pod "45210d52-d52e-4809-b571-ca076d6e9d84" (UID: "45210d52-d52e-4809-b571-ca076d6e9d84"). InnerVolumeSpecName "kube-api-access-v99kb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:34:27 crc kubenswrapper[4923]: I0311 13:34:27.918095 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45210d52-d52e-4809-b571-ca076d6e9d84-util" (OuterVolumeSpecName: "util") pod "45210d52-d52e-4809-b571-ca076d6e9d84" (UID: "45210d52-d52e-4809-b571-ca076d6e9d84"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:34:27 crc kubenswrapper[4923]: I0311 13:34:27.999663 4923 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/45210d52-d52e-4809-b571-ca076d6e9d84-bundle\") on node \"crc\" DevicePath \"\"" Mar 11 13:34:28 crc kubenswrapper[4923]: I0311 13:34:28.000136 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v99kb\" (UniqueName: \"kubernetes.io/projected/45210d52-d52e-4809-b571-ca076d6e9d84-kube-api-access-v99kb\") on node \"crc\" DevicePath \"\"" Mar 11 13:34:28 crc kubenswrapper[4923]: I0311 13:34:28.000379 4923 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/45210d52-d52e-4809-b571-ca076d6e9d84-util\") on node \"crc\" DevicePath \"\"" Mar 11 13:34:28 crc kubenswrapper[4923]: I0311 13:34:28.477194 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww" event={"ID":"45210d52-d52e-4809-b571-ca076d6e9d84","Type":"ContainerDied","Data":"7eb205dc406e5022702720ac1e2018644f6c4e421c743ab5e2c34991323d39b0"} Mar 11 13:34:28 crc kubenswrapper[4923]: I0311 13:34:28.477253 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7eb205dc406e5022702720ac1e2018644f6c4e421c743ab5e2c34991323d39b0" Mar 11 13:34:28 crc kubenswrapper[4923]: I0311 13:34:28.477292 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww" Mar 11 13:34:33 crc kubenswrapper[4923]: I0311 13:34:33.585085 4923 scope.go:117] "RemoveContainer" containerID="46042e4f6d77fb1afb0946f9286c6e087113489c7474adf289f0c1f5aeba6f4f" Mar 11 13:34:36 crc kubenswrapper[4923]: I0311 13:34:36.922949 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-8547b8bf69-qqn9m"] Mar 11 13:34:36 crc kubenswrapper[4923]: E0311 13:34:36.923538 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45210d52-d52e-4809-b571-ca076d6e9d84" containerName="pull" Mar 11 13:34:36 crc kubenswrapper[4923]: I0311 13:34:36.923550 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="45210d52-d52e-4809-b571-ca076d6e9d84" containerName="pull" Mar 11 13:34:36 crc kubenswrapper[4923]: E0311 13:34:36.923569 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45210d52-d52e-4809-b571-ca076d6e9d84" containerName="extract" Mar 11 13:34:36 crc kubenswrapper[4923]: I0311 13:34:36.923576 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="45210d52-d52e-4809-b571-ca076d6e9d84" containerName="extract" Mar 11 13:34:36 crc kubenswrapper[4923]: E0311 13:34:36.923585 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45210d52-d52e-4809-b571-ca076d6e9d84" containerName="util" Mar 11 13:34:36 crc kubenswrapper[4923]: I0311 13:34:36.923592 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="45210d52-d52e-4809-b571-ca076d6e9d84" containerName="util" Mar 11 13:34:36 crc kubenswrapper[4923]: I0311 13:34:36.923690 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="45210d52-d52e-4809-b571-ca076d6e9d84" containerName="extract" Mar 11 13:34:36 crc kubenswrapper[4923]: I0311 13:34:36.924026 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-8547b8bf69-qqn9m" Mar 11 13:34:36 crc kubenswrapper[4923]: I0311 13:34:36.926461 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-k6569" Mar 11 13:34:36 crc kubenswrapper[4923]: I0311 13:34:36.926501 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Mar 11 13:34:36 crc kubenswrapper[4923]: I0311 13:34:36.926831 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Mar 11 13:34:36 crc kubenswrapper[4923]: I0311 13:34:36.927621 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Mar 11 13:34:36 crc kubenswrapper[4923]: I0311 13:34:36.929763 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Mar 11 13:34:36 crc kubenswrapper[4923]: I0311 13:34:36.943064 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-8547b8bf69-qqn9m"] Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.032512 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4be74972-4fdd-48b0-bb91-462c20ecec81-webhook-cert\") pod \"metallb-operator-controller-manager-8547b8bf69-qqn9m\" (UID: \"4be74972-4fdd-48b0-bb91-462c20ecec81\") " pod="metallb-system/metallb-operator-controller-manager-8547b8bf69-qqn9m" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.032560 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4be74972-4fdd-48b0-bb91-462c20ecec81-apiservice-cert\") pod \"metallb-operator-controller-manager-8547b8bf69-qqn9m\" (UID: \"4be74972-4fdd-48b0-bb91-462c20ecec81\") " pod="metallb-system/metallb-operator-controller-manager-8547b8bf69-qqn9m" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.032586 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glf66\" (UniqueName: \"kubernetes.io/projected/4be74972-4fdd-48b0-bb91-462c20ecec81-kube-api-access-glf66\") pod \"metallb-operator-controller-manager-8547b8bf69-qqn9m\" (UID: \"4be74972-4fdd-48b0-bb91-462c20ecec81\") " pod="metallb-system/metallb-operator-controller-manager-8547b8bf69-qqn9m" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.134054 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4be74972-4fdd-48b0-bb91-462c20ecec81-webhook-cert\") pod \"metallb-operator-controller-manager-8547b8bf69-qqn9m\" (UID: \"4be74972-4fdd-48b0-bb91-462c20ecec81\") " pod="metallb-system/metallb-operator-controller-manager-8547b8bf69-qqn9m" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.134106 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4be74972-4fdd-48b0-bb91-462c20ecec81-apiservice-cert\") pod \"metallb-operator-controller-manager-8547b8bf69-qqn9m\" (UID: \"4be74972-4fdd-48b0-bb91-462c20ecec81\") " pod="metallb-system/metallb-operator-controller-manager-8547b8bf69-qqn9m" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.134127 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glf66\" (UniqueName: \"kubernetes.io/projected/4be74972-4fdd-48b0-bb91-462c20ecec81-kube-api-access-glf66\") pod \"metallb-operator-controller-manager-8547b8bf69-qqn9m\" (UID: \"4be74972-4fdd-48b0-bb91-462c20ecec81\") " pod="metallb-system/metallb-operator-controller-manager-8547b8bf69-qqn9m" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.143545 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4be74972-4fdd-48b0-bb91-462c20ecec81-apiservice-cert\") pod \"metallb-operator-controller-manager-8547b8bf69-qqn9m\" (UID: \"4be74972-4fdd-48b0-bb91-462c20ecec81\") " pod="metallb-system/metallb-operator-controller-manager-8547b8bf69-qqn9m" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.152822 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4be74972-4fdd-48b0-bb91-462c20ecec81-webhook-cert\") pod \"metallb-operator-controller-manager-8547b8bf69-qqn9m\" (UID: \"4be74972-4fdd-48b0-bb91-462c20ecec81\") " pod="metallb-system/metallb-operator-controller-manager-8547b8bf69-qqn9m" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.154982 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glf66\" (UniqueName: \"kubernetes.io/projected/4be74972-4fdd-48b0-bb91-462c20ecec81-kube-api-access-glf66\") pod \"metallb-operator-controller-manager-8547b8bf69-qqn9m\" (UID: \"4be74972-4fdd-48b0-bb91-462c20ecec81\") " pod="metallb-system/metallb-operator-controller-manager-8547b8bf69-qqn9m" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.238050 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-8547b8bf69-qqn9m" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.354525 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-5865df787d-xplvf"] Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.355383 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5865df787d-xplvf" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.358008 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.359963 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-66cx9" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.361940 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.373672 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5865df787d-xplvf"] Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.440244 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a56ae696-0c98-464b-a2de-75204772866b-apiservice-cert\") pod \"metallb-operator-webhook-server-5865df787d-xplvf\" (UID: \"a56ae696-0c98-464b-a2de-75204772866b\") " pod="metallb-system/metallb-operator-webhook-server-5865df787d-xplvf" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.440298 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a56ae696-0c98-464b-a2de-75204772866b-webhook-cert\") pod \"metallb-operator-webhook-server-5865df787d-xplvf\" (UID: \"a56ae696-0c98-464b-a2de-75204772866b\") " pod="metallb-system/metallb-operator-webhook-server-5865df787d-xplvf" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.440372 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g8r9\" (UniqueName: \"kubernetes.io/projected/a56ae696-0c98-464b-a2de-75204772866b-kube-api-access-8g8r9\") pod \"metallb-operator-webhook-server-5865df787d-xplvf\" (UID: \"a56ae696-0c98-464b-a2de-75204772866b\") " pod="metallb-system/metallb-operator-webhook-server-5865df787d-xplvf" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.541353 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a56ae696-0c98-464b-a2de-75204772866b-apiservice-cert\") pod \"metallb-operator-webhook-server-5865df787d-xplvf\" (UID: \"a56ae696-0c98-464b-a2de-75204772866b\") " pod="metallb-system/metallb-operator-webhook-server-5865df787d-xplvf" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.541410 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a56ae696-0c98-464b-a2de-75204772866b-webhook-cert\") pod \"metallb-operator-webhook-server-5865df787d-xplvf\" (UID: \"a56ae696-0c98-464b-a2de-75204772866b\") " pod="metallb-system/metallb-operator-webhook-server-5865df787d-xplvf" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.541454 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g8r9\" (UniqueName: \"kubernetes.io/projected/a56ae696-0c98-464b-a2de-75204772866b-kube-api-access-8g8r9\") pod \"metallb-operator-webhook-server-5865df787d-xplvf\" (UID: \"a56ae696-0c98-464b-a2de-75204772866b\") " pod="metallb-system/metallb-operator-webhook-server-5865df787d-xplvf" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.545416 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a56ae696-0c98-464b-a2de-75204772866b-apiservice-cert\") pod \"metallb-operator-webhook-server-5865df787d-xplvf\" (UID: \"a56ae696-0c98-464b-a2de-75204772866b\") " pod="metallb-system/metallb-operator-webhook-server-5865df787d-xplvf" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.545445 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a56ae696-0c98-464b-a2de-75204772866b-webhook-cert\") pod \"metallb-operator-webhook-server-5865df787d-xplvf\" (UID: \"a56ae696-0c98-464b-a2de-75204772866b\") " pod="metallb-system/metallb-operator-webhook-server-5865df787d-xplvf" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.559383 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g8r9\" (UniqueName: \"kubernetes.io/projected/a56ae696-0c98-464b-a2de-75204772866b-kube-api-access-8g8r9\") pod \"metallb-operator-webhook-server-5865df787d-xplvf\" (UID: \"a56ae696-0c98-464b-a2de-75204772866b\") " pod="metallb-system/metallb-operator-webhook-server-5865df787d-xplvf" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.701635 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-5865df787d-xplvf" Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.747254 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-8547b8bf69-qqn9m"] Mar 11 13:34:37 crc kubenswrapper[4923]: W0311 13:34:37.756568 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4be74972_4fdd_48b0_bb91_462c20ecec81.slice/crio-385625fc3852040b0f643ba85abb473f6b46d52a3054cb12b42954436d81d0b6 WatchSource:0}: Error finding container 385625fc3852040b0f643ba85abb473f6b46d52a3054cb12b42954436d81d0b6: Status 404 returned error can't find the container with id 385625fc3852040b0f643ba85abb473f6b46d52a3054cb12b42954436d81d0b6 Mar 11 13:34:37 crc kubenswrapper[4923]: I0311 13:34:37.915665 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-5865df787d-xplvf"] Mar 11 13:34:37 crc kubenswrapper[4923]: W0311 13:34:37.924799 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda56ae696_0c98_464b_a2de_75204772866b.slice/crio-b13f7b031f5e691732c4e51e3302174d019384559e22f49c9d3396a315482231 WatchSource:0}: Error finding container b13f7b031f5e691732c4e51e3302174d019384559e22f49c9d3396a315482231: Status 404 returned error can't find the container with id b13f7b031f5e691732c4e51e3302174d019384559e22f49c9d3396a315482231 Mar 11 13:34:38 crc kubenswrapper[4923]: I0311 13:34:38.532201 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5865df787d-xplvf" event={"ID":"a56ae696-0c98-464b-a2de-75204772866b","Type":"ContainerStarted","Data":"b13f7b031f5e691732c4e51e3302174d019384559e22f49c9d3396a315482231"} Mar 11 13:34:38 crc kubenswrapper[4923]: I0311 13:34:38.533620 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-8547b8bf69-qqn9m" event={"ID":"4be74972-4fdd-48b0-bb91-462c20ecec81","Type":"ContainerStarted","Data":"385625fc3852040b0f643ba85abb473f6b46d52a3054cb12b42954436d81d0b6"} Mar 11 13:34:40 crc kubenswrapper[4923]: I0311 13:34:40.444815 4923 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 11 13:34:41 crc kubenswrapper[4923]: I0311 13:34:41.932513 4923 patch_prober.go:28] interesting pod/machine-config-daemon-fsfnw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 11 13:34:41 crc kubenswrapper[4923]: I0311 13:34:41.932887 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 11 13:34:41 crc kubenswrapper[4923]: I0311 13:34:41.932930 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" Mar 11 13:34:41 crc kubenswrapper[4923]: I0311 13:34:41.933495 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e9d14dc9fced1f5ac360fb70b07bfe3d2a670f2c82a614a9b6443f96729aaa50"} pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 11 13:34:41 crc kubenswrapper[4923]: I0311 13:34:41.933551 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" containerID="cri-o://e9d14dc9fced1f5ac360fb70b07bfe3d2a670f2c82a614a9b6443f96729aaa50" gracePeriod=600 Mar 11 13:34:42 crc kubenswrapper[4923]: I0311 13:34:42.565668 4923 generic.go:334] "Generic (PLEG): container finished" podID="356de301-ded3-4112-8406-1a28e44c2082" containerID="e9d14dc9fced1f5ac360fb70b07bfe3d2a670f2c82a614a9b6443f96729aaa50" exitCode=0 Mar 11 13:34:42 crc kubenswrapper[4923]: I0311 13:34:42.565715 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" event={"ID":"356de301-ded3-4112-8406-1a28e44c2082","Type":"ContainerDied","Data":"e9d14dc9fced1f5ac360fb70b07bfe3d2a670f2c82a614a9b6443f96729aaa50"} Mar 11 13:34:42 crc kubenswrapper[4923]: I0311 13:34:42.565750 4923 scope.go:117] "RemoveContainer" containerID="e24d6b8a7efc75b0eaec16ef2d2e5b49a3a73ee933ce1436580a5243233d6453" Mar 11 13:34:42 crc kubenswrapper[4923]: I0311 13:34:42.798795 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jr57t"] Mar 11 13:34:42 crc kubenswrapper[4923]: I0311 13:34:42.800510 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jr57t" Mar 11 13:34:42 crc kubenswrapper[4923]: I0311 13:34:42.805853 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jr57t"] Mar 11 13:34:42 crc kubenswrapper[4923]: I0311 13:34:42.817772 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08d2747b-a847-4d94-a5cd-6aa9f335100d-utilities\") pod \"certified-operators-jr57t\" (UID: \"08d2747b-a847-4d94-a5cd-6aa9f335100d\") " pod="openshift-marketplace/certified-operators-jr57t" Mar 11 13:34:42 crc kubenswrapper[4923]: I0311 13:34:42.817812 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcs6b\" (UniqueName: \"kubernetes.io/projected/08d2747b-a847-4d94-a5cd-6aa9f335100d-kube-api-access-zcs6b\") pod \"certified-operators-jr57t\" (UID: \"08d2747b-a847-4d94-a5cd-6aa9f335100d\") " pod="openshift-marketplace/certified-operators-jr57t" Mar 11 13:34:42 crc kubenswrapper[4923]: I0311 13:34:42.817872 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08d2747b-a847-4d94-a5cd-6aa9f335100d-catalog-content\") pod \"certified-operators-jr57t\" (UID: \"08d2747b-a847-4d94-a5cd-6aa9f335100d\") " pod="openshift-marketplace/certified-operators-jr57t" Mar 11 13:34:42 crc kubenswrapper[4923]: I0311 13:34:42.918449 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08d2747b-a847-4d94-a5cd-6aa9f335100d-catalog-content\") pod \"certified-operators-jr57t\" (UID: \"08d2747b-a847-4d94-a5cd-6aa9f335100d\") " pod="openshift-marketplace/certified-operators-jr57t" Mar 11 13:34:42 crc kubenswrapper[4923]: I0311 13:34:42.918527 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08d2747b-a847-4d94-a5cd-6aa9f335100d-utilities\") pod \"certified-operators-jr57t\" (UID: \"08d2747b-a847-4d94-a5cd-6aa9f335100d\") " pod="openshift-marketplace/certified-operators-jr57t" Mar 11 13:34:42 crc kubenswrapper[4923]: I0311 13:34:42.918546 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcs6b\" (UniqueName: \"kubernetes.io/projected/08d2747b-a847-4d94-a5cd-6aa9f335100d-kube-api-access-zcs6b\") pod \"certified-operators-jr57t\" (UID: \"08d2747b-a847-4d94-a5cd-6aa9f335100d\") " pod="openshift-marketplace/certified-operators-jr57t" Mar 11 13:34:42 crc kubenswrapper[4923]: I0311 13:34:42.918962 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08d2747b-a847-4d94-a5cd-6aa9f335100d-utilities\") pod \"certified-operators-jr57t\" (UID: \"08d2747b-a847-4d94-a5cd-6aa9f335100d\") " pod="openshift-marketplace/certified-operators-jr57t" Mar 11 13:34:42 crc kubenswrapper[4923]: I0311 13:34:42.919299 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08d2747b-a847-4d94-a5cd-6aa9f335100d-catalog-content\") pod \"certified-operators-jr57t\" (UID: \"08d2747b-a847-4d94-a5cd-6aa9f335100d\") " pod="openshift-marketplace/certified-operators-jr57t" Mar 11 13:34:42 crc kubenswrapper[4923]: I0311 13:34:42.940530 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcs6b\" (UniqueName: \"kubernetes.io/projected/08d2747b-a847-4d94-a5cd-6aa9f335100d-kube-api-access-zcs6b\") pod \"certified-operators-jr57t\" (UID: \"08d2747b-a847-4d94-a5cd-6aa9f335100d\") " pod="openshift-marketplace/certified-operators-jr57t" Mar 11 13:34:43 crc kubenswrapper[4923]: I0311 13:34:43.124042 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jr57t" Mar 11 13:34:43 crc kubenswrapper[4923]: I0311 13:34:43.336224 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jr57t"] Mar 11 13:34:43 crc kubenswrapper[4923]: I0311 13:34:43.573494 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-8547b8bf69-qqn9m" event={"ID":"4be74972-4fdd-48b0-bb91-462c20ecec81","Type":"ContainerStarted","Data":"963a7c94ecc9a52ed0f7939ab5ee62292c75a2d48cdf64078cc58cf32e9a0318"} Mar 11 13:34:43 crc kubenswrapper[4923]: I0311 13:34:43.573579 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-8547b8bf69-qqn9m" Mar 11 13:34:43 crc kubenswrapper[4923]: I0311 13:34:43.575117 4923 generic.go:334] "Generic (PLEG): container finished" podID="08d2747b-a847-4d94-a5cd-6aa9f335100d" containerID="88e6df0a384cfd6c7c401d17d3b82f9078e95c8390ce7bee5d24ce218eeb5f77" exitCode=0 Mar 11 13:34:43 crc kubenswrapper[4923]: I0311 13:34:43.575144 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr57t" event={"ID":"08d2747b-a847-4d94-a5cd-6aa9f335100d","Type":"ContainerDied","Data":"88e6df0a384cfd6c7c401d17d3b82f9078e95c8390ce7bee5d24ce218eeb5f77"} Mar 11 13:34:43 crc kubenswrapper[4923]: I0311 13:34:43.575173 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr57t" event={"ID":"08d2747b-a847-4d94-a5cd-6aa9f335100d","Type":"ContainerStarted","Data":"5f9c32fedfee2d0827d48b562ed0c340ae731b625a1846f7835a49cc030351e4"} Mar 11 13:34:43 crc kubenswrapper[4923]: I0311 13:34:43.576967 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-5865df787d-xplvf" event={"ID":"a56ae696-0c98-464b-a2de-75204772866b","Type":"ContainerStarted","Data":"bf7db34a9866dcc04999f31ebf6dcd21d86e98345a7965e0c041ce1e10c4f126"} Mar 11 13:34:43 crc kubenswrapper[4923]: I0311 13:34:43.577038 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-5865df787d-xplvf" Mar 11 13:34:43 crc kubenswrapper[4923]: I0311 13:34:43.579048 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" event={"ID":"356de301-ded3-4112-8406-1a28e44c2082","Type":"ContainerStarted","Data":"a555ce4420a87a2fc08d149d6004aea3fd8a76f007c748edd82c76ca49ddb60a"} Mar 11 13:34:43 crc kubenswrapper[4923]: I0311 13:34:43.610002 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-8547b8bf69-qqn9m" podStartSLOduration=2.371630685 podStartE2EDuration="7.609985591s" podCreationTimestamp="2026-03-11 13:34:36 +0000 UTC" firstStartedPulling="2026-03-11 13:34:37.758654924 +0000 UTC m=+725.440556648" lastFinishedPulling="2026-03-11 13:34:42.99700984 +0000 UTC m=+730.678911554" observedRunningTime="2026-03-11 13:34:43.605544656 +0000 UTC m=+731.287446370" watchObservedRunningTime="2026-03-11 13:34:43.609985591 +0000 UTC m=+731.291887305" Mar 11 13:34:43 crc kubenswrapper[4923]: I0311 13:34:43.673923 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-5865df787d-xplvf" podStartSLOduration=1.5847282630000001 podStartE2EDuration="6.67390661s" podCreationTimestamp="2026-03-11 13:34:37 +0000 UTC" firstStartedPulling="2026-03-11 13:34:37.928668299 +0000 UTC m=+725.610570013" lastFinishedPulling="2026-03-11 13:34:43.017846646 +0000 UTC m=+730.699748360" observedRunningTime="2026-03-11 13:34:43.672367177 +0000 UTC m=+731.354268891" watchObservedRunningTime="2026-03-11 13:34:43.67390661 +0000 UTC m=+731.355808324" Mar 11 13:34:48 crc kubenswrapper[4923]: I0311 13:34:48.785699 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t9vnp"] Mar 11 13:34:48 crc kubenswrapper[4923]: I0311 13:34:48.787358 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t9vnp" Mar 11 13:34:48 crc kubenswrapper[4923]: I0311 13:34:48.798695 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t9vnp"] Mar 11 13:34:48 crc kubenswrapper[4923]: I0311 13:34:48.898002 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54txw\" (UniqueName: \"kubernetes.io/projected/0850c5c1-23b9-4998-b29f-1c5ab56b8866-kube-api-access-54txw\") pod \"community-operators-t9vnp\" (UID: \"0850c5c1-23b9-4998-b29f-1c5ab56b8866\") " pod="openshift-marketplace/community-operators-t9vnp" Mar 11 13:34:48 crc kubenswrapper[4923]: I0311 13:34:48.898076 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0850c5c1-23b9-4998-b29f-1c5ab56b8866-utilities\") pod \"community-operators-t9vnp\" (UID: \"0850c5c1-23b9-4998-b29f-1c5ab56b8866\") " pod="openshift-marketplace/community-operators-t9vnp" Mar 11 13:34:48 crc kubenswrapper[4923]: I0311 13:34:48.898124 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0850c5c1-23b9-4998-b29f-1c5ab56b8866-catalog-content\") pod \"community-operators-t9vnp\" (UID: \"0850c5c1-23b9-4998-b29f-1c5ab56b8866\") " pod="openshift-marketplace/community-operators-t9vnp" Mar 11 13:34:48 crc kubenswrapper[4923]: I0311 13:34:48.999639 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54txw\" (UniqueName: \"kubernetes.io/projected/0850c5c1-23b9-4998-b29f-1c5ab56b8866-kube-api-access-54txw\") pod \"community-operators-t9vnp\" (UID: \"0850c5c1-23b9-4998-b29f-1c5ab56b8866\") " pod="openshift-marketplace/community-operators-t9vnp" Mar 11 13:34:48 crc kubenswrapper[4923]: I0311 13:34:48.999706 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0850c5c1-23b9-4998-b29f-1c5ab56b8866-utilities\") pod \"community-operators-t9vnp\" (UID: \"0850c5c1-23b9-4998-b29f-1c5ab56b8866\") " pod="openshift-marketplace/community-operators-t9vnp" Mar 11 13:34:48 crc kubenswrapper[4923]: I0311 13:34:48.999754 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0850c5c1-23b9-4998-b29f-1c5ab56b8866-catalog-content\") pod \"community-operators-t9vnp\" (UID: \"0850c5c1-23b9-4998-b29f-1c5ab56b8866\") " pod="openshift-marketplace/community-operators-t9vnp" Mar 11 13:34:49 crc kubenswrapper[4923]: I0311 13:34:49.000616 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0850c5c1-23b9-4998-b29f-1c5ab56b8866-catalog-content\") pod \"community-operators-t9vnp\" (UID: \"0850c5c1-23b9-4998-b29f-1c5ab56b8866\") " pod="openshift-marketplace/community-operators-t9vnp" Mar 11 13:34:49 crc kubenswrapper[4923]: I0311 13:34:49.001427 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0850c5c1-23b9-4998-b29f-1c5ab56b8866-utilities\") pod \"community-operators-t9vnp\" (UID: \"0850c5c1-23b9-4998-b29f-1c5ab56b8866\") " pod="openshift-marketplace/community-operators-t9vnp" Mar 11 13:34:49 crc kubenswrapper[4923]: I0311 13:34:49.027552 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54txw\" (UniqueName: \"kubernetes.io/projected/0850c5c1-23b9-4998-b29f-1c5ab56b8866-kube-api-access-54txw\") pod \"community-operators-t9vnp\" (UID: \"0850c5c1-23b9-4998-b29f-1c5ab56b8866\") " pod="openshift-marketplace/community-operators-t9vnp" Mar 11 13:34:49 crc kubenswrapper[4923]: I0311 13:34:49.107889 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t9vnp" Mar 11 13:34:49 crc kubenswrapper[4923]: I0311 13:34:49.541524 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t9vnp"] Mar 11 13:34:49 crc kubenswrapper[4923]: W0311 13:34:49.554412 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0850c5c1_23b9_4998_b29f_1c5ab56b8866.slice/crio-78b092d928c0a07063b79464c270eda9bc0391e54a284833d34f33f905142898 WatchSource:0}: Error finding container 78b092d928c0a07063b79464c270eda9bc0391e54a284833d34f33f905142898: Status 404 returned error can't find the container with id 78b092d928c0a07063b79464c270eda9bc0391e54a284833d34f33f905142898 Mar 11 13:34:49 crc kubenswrapper[4923]: I0311 13:34:49.612103 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr57t" event={"ID":"08d2747b-a847-4d94-a5cd-6aa9f335100d","Type":"ContainerStarted","Data":"1f5e2caaa97a8a86f1292de9134c79951cd8aa6a72323126a3758789cad88887"} Mar 11 13:34:49 crc kubenswrapper[4923]: I0311 13:34:49.613587 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9vnp" event={"ID":"0850c5c1-23b9-4998-b29f-1c5ab56b8866","Type":"ContainerStarted","Data":"78b092d928c0a07063b79464c270eda9bc0391e54a284833d34f33f905142898"} Mar 11 13:34:50 crc kubenswrapper[4923]: I0311 13:34:50.623280 4923 generic.go:334] "Generic (PLEG): container finished" podID="08d2747b-a847-4d94-a5cd-6aa9f335100d" containerID="1f5e2caaa97a8a86f1292de9134c79951cd8aa6a72323126a3758789cad88887" exitCode=0 Mar 11 13:34:50 crc kubenswrapper[4923]: I0311 13:34:50.623424 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr57t" event={"ID":"08d2747b-a847-4d94-a5cd-6aa9f335100d","Type":"ContainerDied","Data":"1f5e2caaa97a8a86f1292de9134c79951cd8aa6a72323126a3758789cad88887"} Mar 11 13:34:50 crc kubenswrapper[4923]: I0311 13:34:50.625377 4923 generic.go:334] "Generic (PLEG): container finished" podID="0850c5c1-23b9-4998-b29f-1c5ab56b8866" containerID="3891f3e108a17204791c72952e4d5b57e05884f9bd8221b1f44bda12da9657ca" exitCode=0 Mar 11 13:34:50 crc kubenswrapper[4923]: I0311 13:34:50.625415 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9vnp" event={"ID":"0850c5c1-23b9-4998-b29f-1c5ab56b8866","Type":"ContainerDied","Data":"3891f3e108a17204791c72952e4d5b57e05884f9bd8221b1f44bda12da9657ca"} Mar 11 13:34:51 crc kubenswrapper[4923]: I0311 13:34:51.646917 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr57t" event={"ID":"08d2747b-a847-4d94-a5cd-6aa9f335100d","Type":"ContainerStarted","Data":"656a171a79f77d113a8858f739a2df4c38e28e8333a2b816604cd74833127992"} Mar 11 13:34:51 crc kubenswrapper[4923]: I0311 13:34:51.681675 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jr57t" podStartSLOduration=2.174314904 podStartE2EDuration="9.681652177s" podCreationTimestamp="2026-03-11 13:34:42 +0000 UTC" firstStartedPulling="2026-03-11 13:34:43.576309094 +0000 UTC m=+731.258210808" lastFinishedPulling="2026-03-11 13:34:51.083646367 +0000 UTC m=+738.765548081" observedRunningTime="2026-03-11 13:34:51.673624941 +0000 UTC m=+739.355526675" watchObservedRunningTime="2026-03-11 13:34:51.681652177 +0000 UTC m=+739.363553901" Mar 11 13:34:52 crc kubenswrapper[4923]: I0311 13:34:52.660771 4923 generic.go:334] "Generic (PLEG): container finished" podID="0850c5c1-23b9-4998-b29f-1c5ab56b8866" containerID="ef55aac659520c8c262754886694efc896981cbcad5c3b1e4eea87ad8c2ab545" exitCode=0 Mar 11 13:34:52 crc kubenswrapper[4923]: I0311 13:34:52.661593 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9vnp" event={"ID":"0850c5c1-23b9-4998-b29f-1c5ab56b8866","Type":"ContainerDied","Data":"ef55aac659520c8c262754886694efc896981cbcad5c3b1e4eea87ad8c2ab545"} Mar 11 13:34:53 crc kubenswrapper[4923]: I0311 13:34:53.124738 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jr57t" Mar 11 13:34:53 crc kubenswrapper[4923]: I0311 13:34:53.125184 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jr57t" Mar 11 13:34:53 crc kubenswrapper[4923]: I0311 13:34:53.668128 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9vnp" event={"ID":"0850c5c1-23b9-4998-b29f-1c5ab56b8866","Type":"ContainerStarted","Data":"80296d93d091d4baeb9394a9fcc3c498dcf45eb4fcc7b97e551021de2888f5f9"} Mar 11 13:34:53 crc kubenswrapper[4923]: I0311 13:34:53.685144 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t9vnp" podStartSLOduration=3.220012393 podStartE2EDuration="5.685130391s" podCreationTimestamp="2026-03-11 13:34:48 +0000 UTC" firstStartedPulling="2026-03-11 13:34:50.626784428 +0000 UTC m=+738.308686162" lastFinishedPulling="2026-03-11 13:34:53.091902406 +0000 UTC m=+740.773804160" observedRunningTime="2026-03-11 13:34:53.683050543 +0000 UTC m=+741.364952257" watchObservedRunningTime="2026-03-11 13:34:53.685130391 +0000 UTC m=+741.367032105" Mar 11 13:34:54 crc kubenswrapper[4923]: I0311 13:34:54.183186 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-jr57t" podUID="08d2747b-a847-4d94-a5cd-6aa9f335100d" containerName="registry-server" probeResult="failure" output=< Mar 11 13:34:54 crc kubenswrapper[4923]: timeout: failed to connect service ":50051" within 1s Mar 11 13:34:54 crc kubenswrapper[4923]: > Mar 11 13:34:57 crc kubenswrapper[4923]: I0311 13:34:57.713361 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-5865df787d-xplvf" Mar 11 13:34:59 crc kubenswrapper[4923]: I0311 13:34:59.109246 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t9vnp" Mar 11 13:34:59 crc kubenswrapper[4923]: I0311 13:34:59.109747 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t9vnp" Mar 11 13:34:59 crc kubenswrapper[4923]: I0311 13:34:59.161037 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t9vnp" Mar 11 13:34:59 crc kubenswrapper[4923]: I0311 13:34:59.756319 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t9vnp" Mar 11 13:35:00 crc kubenswrapper[4923]: I0311 13:35:00.776623 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t9vnp"] Mar 11 13:35:01 crc kubenswrapper[4923]: I0311 13:35:01.715870 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t9vnp" podUID="0850c5c1-23b9-4998-b29f-1c5ab56b8866" containerName="registry-server" containerID="cri-o://80296d93d091d4baeb9394a9fcc3c498dcf45eb4fcc7b97e551021de2888f5f9" gracePeriod=2 Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.065450 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t9vnp" Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.174826 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0850c5c1-23b9-4998-b29f-1c5ab56b8866-catalog-content\") pod \"0850c5c1-23b9-4998-b29f-1c5ab56b8866\" (UID: \"0850c5c1-23b9-4998-b29f-1c5ab56b8866\") " Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.174897 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0850c5c1-23b9-4998-b29f-1c5ab56b8866-utilities\") pod \"0850c5c1-23b9-4998-b29f-1c5ab56b8866\" (UID: \"0850c5c1-23b9-4998-b29f-1c5ab56b8866\") " Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.175019 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54txw\" (UniqueName: \"kubernetes.io/projected/0850c5c1-23b9-4998-b29f-1c5ab56b8866-kube-api-access-54txw\") pod \"0850c5c1-23b9-4998-b29f-1c5ab56b8866\" (UID: \"0850c5c1-23b9-4998-b29f-1c5ab56b8866\") " Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.175915 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0850c5c1-23b9-4998-b29f-1c5ab56b8866-utilities" (OuterVolumeSpecName: "utilities") pod "0850c5c1-23b9-4998-b29f-1c5ab56b8866" (UID: "0850c5c1-23b9-4998-b29f-1c5ab56b8866"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.180385 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0850c5c1-23b9-4998-b29f-1c5ab56b8866-kube-api-access-54txw" (OuterVolumeSpecName: "kube-api-access-54txw") pod "0850c5c1-23b9-4998-b29f-1c5ab56b8866" (UID: "0850c5c1-23b9-4998-b29f-1c5ab56b8866"). InnerVolumeSpecName "kube-api-access-54txw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.236945 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0850c5c1-23b9-4998-b29f-1c5ab56b8866-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0850c5c1-23b9-4998-b29f-1c5ab56b8866" (UID: "0850c5c1-23b9-4998-b29f-1c5ab56b8866"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.275950 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0850c5c1-23b9-4998-b29f-1c5ab56b8866-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.276032 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0850c5c1-23b9-4998-b29f-1c5ab56b8866-utilities\") on node \"crc\" DevicePath \"\"" Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.276052 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54txw\" (UniqueName: \"kubernetes.io/projected/0850c5c1-23b9-4998-b29f-1c5ab56b8866-kube-api-access-54txw\") on node \"crc\" DevicePath \"\"" Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.725252 4923 generic.go:334] "Generic (PLEG): container finished" podID="0850c5c1-23b9-4998-b29f-1c5ab56b8866" containerID="80296d93d091d4baeb9394a9fcc3c498dcf45eb4fcc7b97e551021de2888f5f9" exitCode=0 Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.725303 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9vnp" event={"ID":"0850c5c1-23b9-4998-b29f-1c5ab56b8866","Type":"ContainerDied","Data":"80296d93d091d4baeb9394a9fcc3c498dcf45eb4fcc7b97e551021de2888f5f9"} Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.725306 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t9vnp" Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.725334 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t9vnp" event={"ID":"0850c5c1-23b9-4998-b29f-1c5ab56b8866","Type":"ContainerDied","Data":"78b092d928c0a07063b79464c270eda9bc0391e54a284833d34f33f905142898"} Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.725392 4923 scope.go:117] "RemoveContainer" containerID="80296d93d091d4baeb9394a9fcc3c498dcf45eb4fcc7b97e551021de2888f5f9" Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.745806 4923 scope.go:117] "RemoveContainer" containerID="ef55aac659520c8c262754886694efc896981cbcad5c3b1e4eea87ad8c2ab545" Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.758631 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t9vnp"] Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.762138 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t9vnp"] Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.783068 4923 scope.go:117] "RemoveContainer" containerID="3891f3e108a17204791c72952e4d5b57e05884f9bd8221b1f44bda12da9657ca" Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.802740 4923 scope.go:117] "RemoveContainer" containerID="80296d93d091d4baeb9394a9fcc3c498dcf45eb4fcc7b97e551021de2888f5f9" Mar 11 13:35:02 crc kubenswrapper[4923]: E0311 13:35:02.803311 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80296d93d091d4baeb9394a9fcc3c498dcf45eb4fcc7b97e551021de2888f5f9\": container with ID starting with 80296d93d091d4baeb9394a9fcc3c498dcf45eb4fcc7b97e551021de2888f5f9 not found: ID does not exist" containerID="80296d93d091d4baeb9394a9fcc3c498dcf45eb4fcc7b97e551021de2888f5f9" Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.803391 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80296d93d091d4baeb9394a9fcc3c498dcf45eb4fcc7b97e551021de2888f5f9"} err="failed to get container status \"80296d93d091d4baeb9394a9fcc3c498dcf45eb4fcc7b97e551021de2888f5f9\": rpc error: code = NotFound desc = could not find container \"80296d93d091d4baeb9394a9fcc3c498dcf45eb4fcc7b97e551021de2888f5f9\": container with ID starting with 80296d93d091d4baeb9394a9fcc3c498dcf45eb4fcc7b97e551021de2888f5f9 not found: ID does not exist" Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.803425 4923 scope.go:117] "RemoveContainer" containerID="ef55aac659520c8c262754886694efc896981cbcad5c3b1e4eea87ad8c2ab545" Mar 11 13:35:02 crc kubenswrapper[4923]: E0311 13:35:02.803895 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef55aac659520c8c262754886694efc896981cbcad5c3b1e4eea87ad8c2ab545\": container with ID starting with ef55aac659520c8c262754886694efc896981cbcad5c3b1e4eea87ad8c2ab545 not found: ID does not exist" containerID="ef55aac659520c8c262754886694efc896981cbcad5c3b1e4eea87ad8c2ab545" Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.803938 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef55aac659520c8c262754886694efc896981cbcad5c3b1e4eea87ad8c2ab545"} err="failed to get container status \"ef55aac659520c8c262754886694efc896981cbcad5c3b1e4eea87ad8c2ab545\": rpc error: code = NotFound desc = could not find container \"ef55aac659520c8c262754886694efc896981cbcad5c3b1e4eea87ad8c2ab545\": container with ID starting with ef55aac659520c8c262754886694efc896981cbcad5c3b1e4eea87ad8c2ab545 not found: ID does not exist" Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.803966 4923 scope.go:117] "RemoveContainer" containerID="3891f3e108a17204791c72952e4d5b57e05884f9bd8221b1f44bda12da9657ca" Mar 11 13:35:02 crc kubenswrapper[4923]: E0311 13:35:02.804204 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3891f3e108a17204791c72952e4d5b57e05884f9bd8221b1f44bda12da9657ca\": container with ID starting with 3891f3e108a17204791c72952e4d5b57e05884f9bd8221b1f44bda12da9657ca not found: ID does not exist" containerID="3891f3e108a17204791c72952e4d5b57e05884f9bd8221b1f44bda12da9657ca" Mar 11 13:35:02 crc kubenswrapper[4923]: I0311 13:35:02.804242 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3891f3e108a17204791c72952e4d5b57e05884f9bd8221b1f44bda12da9657ca"} err="failed to get container status \"3891f3e108a17204791c72952e4d5b57e05884f9bd8221b1f44bda12da9657ca\": rpc error: code = NotFound desc = could not find container \"3891f3e108a17204791c72952e4d5b57e05884f9bd8221b1f44bda12da9657ca\": container with ID starting with 3891f3e108a17204791c72952e4d5b57e05884f9bd8221b1f44bda12da9657ca not found: ID does not exist" Mar 11 13:35:03 crc kubenswrapper[4923]: I0311 13:35:03.029197 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0850c5c1-23b9-4998-b29f-1c5ab56b8866" path="/var/lib/kubelet/pods/0850c5c1-23b9-4998-b29f-1c5ab56b8866/volumes" Mar 11 13:35:03 crc kubenswrapper[4923]: I0311 13:35:03.189888 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jr57t" Mar 11 13:35:03 crc kubenswrapper[4923]: I0311 13:35:03.252831 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jr57t" Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.180937 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jr57t"] Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.181656 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jr57t" podUID="08d2747b-a847-4d94-a5cd-6aa9f335100d" containerName="registry-server" containerID="cri-o://656a171a79f77d113a8858f739a2df4c38e28e8333a2b816604cd74833127992" gracePeriod=2 Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.598219 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jr57t" Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.734511 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08d2747b-a847-4d94-a5cd-6aa9f335100d-catalog-content\") pod \"08d2747b-a847-4d94-a5cd-6aa9f335100d\" (UID: \"08d2747b-a847-4d94-a5cd-6aa9f335100d\") " Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.734626 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08d2747b-a847-4d94-a5cd-6aa9f335100d-utilities\") pod \"08d2747b-a847-4d94-a5cd-6aa9f335100d\" (UID: \"08d2747b-a847-4d94-a5cd-6aa9f335100d\") " Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.734672 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zcs6b\" (UniqueName: \"kubernetes.io/projected/08d2747b-a847-4d94-a5cd-6aa9f335100d-kube-api-access-zcs6b\") pod \"08d2747b-a847-4d94-a5cd-6aa9f335100d\" (UID: \"08d2747b-a847-4d94-a5cd-6aa9f335100d\") " Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.736403 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08d2747b-a847-4d94-a5cd-6aa9f335100d-utilities" (OuterVolumeSpecName: "utilities") pod "08d2747b-a847-4d94-a5cd-6aa9f335100d" (UID: "08d2747b-a847-4d94-a5cd-6aa9f335100d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.745131 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08d2747b-a847-4d94-a5cd-6aa9f335100d-kube-api-access-zcs6b" (OuterVolumeSpecName: "kube-api-access-zcs6b") pod "08d2747b-a847-4d94-a5cd-6aa9f335100d" (UID: "08d2747b-a847-4d94-a5cd-6aa9f335100d"). InnerVolumeSpecName "kube-api-access-zcs6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.762811 4923 generic.go:334] "Generic (PLEG): container finished" podID="08d2747b-a847-4d94-a5cd-6aa9f335100d" containerID="656a171a79f77d113a8858f739a2df4c38e28e8333a2b816604cd74833127992" exitCode=0 Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.762880 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr57t" event={"ID":"08d2747b-a847-4d94-a5cd-6aa9f335100d","Type":"ContainerDied","Data":"656a171a79f77d113a8858f739a2df4c38e28e8333a2b816604cd74833127992"} Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.762922 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jr57t" event={"ID":"08d2747b-a847-4d94-a5cd-6aa9f335100d","Type":"ContainerDied","Data":"5f9c32fedfee2d0827d48b562ed0c340ae731b625a1846f7835a49cc030351e4"} Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.762953 4923 scope.go:117] "RemoveContainer" containerID="656a171a79f77d113a8858f739a2df4c38e28e8333a2b816604cd74833127992" Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.763125 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jr57t" Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.792652 4923 scope.go:117] "RemoveContainer" containerID="1f5e2caaa97a8a86f1292de9134c79951cd8aa6a72323126a3758789cad88887" Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.835915 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08d2747b-a847-4d94-a5cd-6aa9f335100d-utilities\") on node \"crc\" DevicePath \"\"" Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.835988 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zcs6b\" (UniqueName: \"kubernetes.io/projected/08d2747b-a847-4d94-a5cd-6aa9f335100d-kube-api-access-zcs6b\") on node \"crc\" DevicePath \"\"" Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.838133 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08d2747b-a847-4d94-a5cd-6aa9f335100d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08d2747b-a847-4d94-a5cd-6aa9f335100d" (UID: "08d2747b-a847-4d94-a5cd-6aa9f335100d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.840625 4923 scope.go:117] "RemoveContainer" containerID="88e6df0a384cfd6c7c401d17d3b82f9078e95c8390ce7bee5d24ce218eeb5f77" Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.863648 4923 scope.go:117] "RemoveContainer" containerID="656a171a79f77d113a8858f739a2df4c38e28e8333a2b816604cd74833127992" Mar 11 13:35:06 crc kubenswrapper[4923]: E0311 13:35:06.864316 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"656a171a79f77d113a8858f739a2df4c38e28e8333a2b816604cd74833127992\": container with ID starting with 656a171a79f77d113a8858f739a2df4c38e28e8333a2b816604cd74833127992 not found: ID does not exist" containerID="656a171a79f77d113a8858f739a2df4c38e28e8333a2b816604cd74833127992" Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.864391 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"656a171a79f77d113a8858f739a2df4c38e28e8333a2b816604cd74833127992"} err="failed to get container status \"656a171a79f77d113a8858f739a2df4c38e28e8333a2b816604cd74833127992\": rpc error: code = NotFound desc = could not find container \"656a171a79f77d113a8858f739a2df4c38e28e8333a2b816604cd74833127992\": container with ID starting with 656a171a79f77d113a8858f739a2df4c38e28e8333a2b816604cd74833127992 not found: ID does not exist" Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.864415 4923 scope.go:117] "RemoveContainer" containerID="1f5e2caaa97a8a86f1292de9134c79951cd8aa6a72323126a3758789cad88887" Mar 11 13:35:06 crc kubenswrapper[4923]: E0311 13:35:06.864859 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f5e2caaa97a8a86f1292de9134c79951cd8aa6a72323126a3758789cad88887\": container with ID starting with 1f5e2caaa97a8a86f1292de9134c79951cd8aa6a72323126a3758789cad88887 not found: ID does not exist" containerID="1f5e2caaa97a8a86f1292de9134c79951cd8aa6a72323126a3758789cad88887" Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.864916 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f5e2caaa97a8a86f1292de9134c79951cd8aa6a72323126a3758789cad88887"} err="failed to get container status \"1f5e2caaa97a8a86f1292de9134c79951cd8aa6a72323126a3758789cad88887\": rpc error: code = NotFound desc = could not find container \"1f5e2caaa97a8a86f1292de9134c79951cd8aa6a72323126a3758789cad88887\": container with ID starting with 1f5e2caaa97a8a86f1292de9134c79951cd8aa6a72323126a3758789cad88887 not found: ID does not exist" Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.864937 4923 scope.go:117] "RemoveContainer" containerID="88e6df0a384cfd6c7c401d17d3b82f9078e95c8390ce7bee5d24ce218eeb5f77" Mar 11 13:35:06 crc kubenswrapper[4923]: E0311 13:35:06.865380 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88e6df0a384cfd6c7c401d17d3b82f9078e95c8390ce7bee5d24ce218eeb5f77\": container with ID starting with 88e6df0a384cfd6c7c401d17d3b82f9078e95c8390ce7bee5d24ce218eeb5f77 not found: ID does not exist" containerID="88e6df0a384cfd6c7c401d17d3b82f9078e95c8390ce7bee5d24ce218eeb5f77" Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.865427 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88e6df0a384cfd6c7c401d17d3b82f9078e95c8390ce7bee5d24ce218eeb5f77"} err="failed to get container status \"88e6df0a384cfd6c7c401d17d3b82f9078e95c8390ce7bee5d24ce218eeb5f77\": rpc error: code = NotFound desc = could not find container \"88e6df0a384cfd6c7c401d17d3b82f9078e95c8390ce7bee5d24ce218eeb5f77\": container with ID starting with 88e6df0a384cfd6c7c401d17d3b82f9078e95c8390ce7bee5d24ce218eeb5f77 not found: ID does not exist" Mar 11 13:35:06 crc kubenswrapper[4923]: I0311 13:35:06.937997 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08d2747b-a847-4d94-a5cd-6aa9f335100d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 11 13:35:07 crc kubenswrapper[4923]: I0311 13:35:07.083701 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jr57t"] Mar 11 13:35:07 crc kubenswrapper[4923]: I0311 13:35:07.088120 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jr57t"] Mar 11 13:35:09 crc kubenswrapper[4923]: I0311 13:35:09.039119 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08d2747b-a847-4d94-a5cd-6aa9f335100d" path="/var/lib/kubelet/pods/08d2747b-a847-4d94-a5cd-6aa9f335100d/volumes" Mar 11 13:35:17 crc kubenswrapper[4923]: I0311 13:35:17.241194 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-8547b8bf69-qqn9m" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.097275 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-zwkrg"] Mar 11 13:35:18 crc kubenswrapper[4923]: E0311 13:35:18.097945 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0850c5c1-23b9-4998-b29f-1c5ab56b8866" containerName="extract-content" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.097970 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="0850c5c1-23b9-4998-b29f-1c5ab56b8866" containerName="extract-content" Mar 11 13:35:18 crc kubenswrapper[4923]: E0311 13:35:18.097986 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0850c5c1-23b9-4998-b29f-1c5ab56b8866" containerName="extract-utilities" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.097997 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="0850c5c1-23b9-4998-b29f-1c5ab56b8866" containerName="extract-utilities" Mar 11 13:35:18 crc kubenswrapper[4923]: E0311 13:35:18.098010 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08d2747b-a847-4d94-a5cd-6aa9f335100d" containerName="extract-utilities" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.098020 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="08d2747b-a847-4d94-a5cd-6aa9f335100d" containerName="extract-utilities" Mar 11 13:35:18 crc kubenswrapper[4923]: E0311 13:35:18.098042 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0850c5c1-23b9-4998-b29f-1c5ab56b8866" containerName="registry-server" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.098052 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="0850c5c1-23b9-4998-b29f-1c5ab56b8866" containerName="registry-server" Mar 11 13:35:18 crc kubenswrapper[4923]: E0311 13:35:18.098069 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08d2747b-a847-4d94-a5cd-6aa9f335100d" containerName="extract-content" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.098078 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="08d2747b-a847-4d94-a5cd-6aa9f335100d" containerName="extract-content" Mar 11 13:35:18 crc kubenswrapper[4923]: E0311 13:35:18.098095 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08d2747b-a847-4d94-a5cd-6aa9f335100d" containerName="registry-server" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.098105 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="08d2747b-a847-4d94-a5cd-6aa9f335100d" containerName="registry-server" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.098332 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="08d2747b-a847-4d94-a5cd-6aa9f335100d" containerName="registry-server" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.098376 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="0850c5c1-23b9-4998-b29f-1c5ab56b8866" containerName="registry-server" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.100134 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-zwkrg" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.102457 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-rrbq7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.103120 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.107531 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-zwkrg"] Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.127035 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-9f8w7"] Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.129262 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.132972 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.133826 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.187259 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0085e43b-b941-4b8b-b0c3-3ca1cdd68c58-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-zwkrg\" (UID: \"0085e43b-b941-4b8b-b0c3-3ca1cdd68c58\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-zwkrg" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.187416 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s62xf\" (UniqueName: \"kubernetes.io/projected/0085e43b-b941-4b8b-b0c3-3ca1cdd68c58-kube-api-access-s62xf\") pod \"frr-k8s-webhook-server-bcc4b6f68-zwkrg\" (UID: \"0085e43b-b941-4b8b-b0c3-3ca1cdd68c58\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-zwkrg" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.241466 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-rrmgz"] Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.242268 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-rrmgz" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.246878 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.246894 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.247100 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-56sqp" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.247189 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.259264 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-7bb4cc7c98-47vwc"] Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.288026 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-47vwc" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.289664 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0-metrics\") pod \"frr-k8s-9f8w7\" (UID: \"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0\") " pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.289700 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0085e43b-b941-4b8b-b0c3-3ca1cdd68c58-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-zwkrg\" (UID: \"0085e43b-b941-4b8b-b0c3-3ca1cdd68c58\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-zwkrg" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.289749 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0-metrics-certs\") pod \"frr-k8s-9f8w7\" (UID: \"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0\") " pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.289770 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmw8v\" (UniqueName: \"kubernetes.io/projected/21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0-kube-api-access-xmw8v\") pod \"frr-k8s-9f8w7\" (UID: \"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0\") " pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.289797 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0-reloader\") pod \"frr-k8s-9f8w7\" (UID: \"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0\") " pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.289827 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0-frr-sockets\") pod \"frr-k8s-9f8w7\" (UID: \"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0\") " pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.289848 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0-frr-conf\") pod \"frr-k8s-9f8w7\" (UID: \"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0\") " pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.289871 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s62xf\" (UniqueName: \"kubernetes.io/projected/0085e43b-b941-4b8b-b0c3-3ca1cdd68c58-kube-api-access-s62xf\") pod \"frr-k8s-webhook-server-bcc4b6f68-zwkrg\" (UID: \"0085e43b-b941-4b8b-b0c3-3ca1cdd68c58\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-zwkrg" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.289905 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0-frr-startup\") pod \"frr-k8s-9f8w7\" (UID: \"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0\") " pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: E0311 13:35:18.290023 4923 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Mar 11 13:35:18 crc kubenswrapper[4923]: E0311 13:35:18.290067 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0085e43b-b941-4b8b-b0c3-3ca1cdd68c58-cert podName:0085e43b-b941-4b8b-b0c3-3ca1cdd68c58 nodeName:}" failed. No retries permitted until 2026-03-11 13:35:18.790050602 +0000 UTC m=+766.471952316 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0085e43b-b941-4b8b-b0c3-3ca1cdd68c58-cert") pod "frr-k8s-webhook-server-bcc4b6f68-zwkrg" (UID: "0085e43b-b941-4b8b-b0c3-3ca1cdd68c58") : secret "frr-k8s-webhook-server-cert" not found Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.291414 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.306575 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-47vwc"] Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.317620 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s62xf\" (UniqueName: \"kubernetes.io/projected/0085e43b-b941-4b8b-b0c3-3ca1cdd68c58-kube-api-access-s62xf\") pod \"frr-k8s-webhook-server-bcc4b6f68-zwkrg\" (UID: \"0085e43b-b941-4b8b-b0c3-3ca1cdd68c58\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-zwkrg" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.390515 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0-frr-sockets\") pod \"frr-k8s-9f8w7\" (UID: \"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0\") " pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.390568 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0-frr-conf\") pod \"frr-k8s-9f8w7\" (UID: \"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0\") " pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.390604 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0-frr-startup\") pod \"frr-k8s-9f8w7\" (UID: \"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0\") " pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.390627 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a4d6dd0e-4763-486d-a6a6-8ce519149aaa-memberlist\") pod \"speaker-rrmgz\" (UID: \"a4d6dd0e-4763-486d-a6a6-8ce519149aaa\") " pod="metallb-system/speaker-rrmgz" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.390646 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0-metrics\") pod \"frr-k8s-9f8w7\" (UID: \"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0\") " pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.390689 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a4d6dd0e-4763-486d-a6a6-8ce519149aaa-metallb-excludel2\") pod \"speaker-rrmgz\" (UID: \"a4d6dd0e-4763-486d-a6a6-8ce519149aaa\") " pod="metallb-system/speaker-rrmgz" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.390736 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0-metrics-certs\") pod \"frr-k8s-9f8w7\" (UID: \"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0\") " pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.390759 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8p7dn\" (UniqueName: \"kubernetes.io/projected/a4d6dd0e-4763-486d-a6a6-8ce519149aaa-kube-api-access-8p7dn\") pod \"speaker-rrmgz\" (UID: \"a4d6dd0e-4763-486d-a6a6-8ce519149aaa\") " pod="metallb-system/speaker-rrmgz" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.390781 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/475686c9-3c06-456e-b570-5b3774a834d9-cert\") pod \"controller-7bb4cc7c98-47vwc\" (UID: \"475686c9-3c06-456e-b570-5b3774a834d9\") " pod="metallb-system/controller-7bb4cc7c98-47vwc" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.390802 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmw8v\" (UniqueName: \"kubernetes.io/projected/21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0-kube-api-access-xmw8v\") pod \"frr-k8s-9f8w7\" (UID: \"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0\") " pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.390819 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4d6dd0e-4763-486d-a6a6-8ce519149aaa-metrics-certs\") pod \"speaker-rrmgz\" (UID: \"a4d6dd0e-4763-486d-a6a6-8ce519149aaa\") " pod="metallb-system/speaker-rrmgz" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.390840 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0-reloader\") pod \"frr-k8s-9f8w7\" (UID: \"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0\") " pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.390862 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/475686c9-3c06-456e-b570-5b3774a834d9-metrics-certs\") pod \"controller-7bb4cc7c98-47vwc\" (UID: \"475686c9-3c06-456e-b570-5b3774a834d9\") " pod="metallb-system/controller-7bb4cc7c98-47vwc" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.390882 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlx6z\" (UniqueName: \"kubernetes.io/projected/475686c9-3c06-456e-b570-5b3774a834d9-kube-api-access-dlx6z\") pod \"controller-7bb4cc7c98-47vwc\" (UID: \"475686c9-3c06-456e-b570-5b3774a834d9\") " pod="metallb-system/controller-7bb4cc7c98-47vwc" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.391078 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0-frr-sockets\") pod \"frr-k8s-9f8w7\" (UID: \"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0\") " pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.391137 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0-frr-conf\") pod \"frr-k8s-9f8w7\" (UID: \"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0\") " pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.391148 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0-reloader\") pod \"frr-k8s-9f8w7\" (UID: \"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0\") " pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.391292 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0-metrics\") pod \"frr-k8s-9f8w7\" (UID: \"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0\") " pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.391859 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0-frr-startup\") pod \"frr-k8s-9f8w7\" (UID: \"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0\") " pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.396036 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0-metrics-certs\") pod \"frr-k8s-9f8w7\" (UID: \"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0\") " pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.413459 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmw8v\" (UniqueName: \"kubernetes.io/projected/21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0-kube-api-access-xmw8v\") pod \"frr-k8s-9f8w7\" (UID: \"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0\") " pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.458001 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.493022 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a4d6dd0e-4763-486d-a6a6-8ce519149aaa-memberlist\") pod \"speaker-rrmgz\" (UID: \"a4d6dd0e-4763-486d-a6a6-8ce519149aaa\") " pod="metallb-system/speaker-rrmgz" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.493302 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a4d6dd0e-4763-486d-a6a6-8ce519149aaa-metallb-excludel2\") pod \"speaker-rrmgz\" (UID: \"a4d6dd0e-4763-486d-a6a6-8ce519149aaa\") " pod="metallb-system/speaker-rrmgz" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.493452 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8p7dn\" (UniqueName: \"kubernetes.io/projected/a4d6dd0e-4763-486d-a6a6-8ce519149aaa-kube-api-access-8p7dn\") pod \"speaker-rrmgz\" (UID: \"a4d6dd0e-4763-486d-a6a6-8ce519149aaa\") " pod="metallb-system/speaker-rrmgz" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.493558 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/475686c9-3c06-456e-b570-5b3774a834d9-cert\") pod \"controller-7bb4cc7c98-47vwc\" (UID: \"475686c9-3c06-456e-b570-5b3774a834d9\") " pod="metallb-system/controller-7bb4cc7c98-47vwc" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.493671 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4d6dd0e-4763-486d-a6a6-8ce519149aaa-metrics-certs\") pod \"speaker-rrmgz\" (UID: \"a4d6dd0e-4763-486d-a6a6-8ce519149aaa\") " pod="metallb-system/speaker-rrmgz" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.493782 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/475686c9-3c06-456e-b570-5b3774a834d9-metrics-certs\") pod \"controller-7bb4cc7c98-47vwc\" (UID: \"475686c9-3c06-456e-b570-5b3774a834d9\") " pod="metallb-system/controller-7bb4cc7c98-47vwc" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.493918 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlx6z\" (UniqueName: \"kubernetes.io/projected/475686c9-3c06-456e-b570-5b3774a834d9-kube-api-access-dlx6z\") pod \"controller-7bb4cc7c98-47vwc\" (UID: \"475686c9-3c06-456e-b570-5b3774a834d9\") " pod="metallb-system/controller-7bb4cc7c98-47vwc" Mar 11 13:35:18 crc kubenswrapper[4923]: E0311 13:35:18.493162 4923 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 11 13:35:18 crc kubenswrapper[4923]: E0311 13:35:18.494151 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4d6dd0e-4763-486d-a6a6-8ce519149aaa-memberlist podName:a4d6dd0e-4763-486d-a6a6-8ce519149aaa nodeName:}" failed. No retries permitted until 2026-03-11 13:35:18.994123876 +0000 UTC m=+766.676025600 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/a4d6dd0e-4763-486d-a6a6-8ce519149aaa-memberlist") pod "speaker-rrmgz" (UID: "a4d6dd0e-4763-486d-a6a6-8ce519149aaa") : secret "metallb-memberlist" not found Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.493980 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a4d6dd0e-4763-486d-a6a6-8ce519149aaa-metallb-excludel2\") pod \"speaker-rrmgz\" (UID: \"a4d6dd0e-4763-486d-a6a6-8ce519149aaa\") " pod="metallb-system/speaker-rrmgz" Mar 11 13:35:18 crc kubenswrapper[4923]: E0311 13:35:18.493783 4923 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Mar 11 13:35:18 crc kubenswrapper[4923]: E0311 13:35:18.494225 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4d6dd0e-4763-486d-a6a6-8ce519149aaa-metrics-certs podName:a4d6dd0e-4763-486d-a6a6-8ce519149aaa nodeName:}" failed. No retries permitted until 2026-03-11 13:35:18.994216578 +0000 UTC m=+766.676118302 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/a4d6dd0e-4763-486d-a6a6-8ce519149aaa-metrics-certs") pod "speaker-rrmgz" (UID: "a4d6dd0e-4763-486d-a6a6-8ce519149aaa") : secret "speaker-certs-secret" not found Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.497368 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/475686c9-3c06-456e-b570-5b3774a834d9-metrics-certs\") pod \"controller-7bb4cc7c98-47vwc\" (UID: \"475686c9-3c06-456e-b570-5b3774a834d9\") " pod="metallb-system/controller-7bb4cc7c98-47vwc" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.498868 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/475686c9-3c06-456e-b570-5b3774a834d9-cert\") pod \"controller-7bb4cc7c98-47vwc\" (UID: \"475686c9-3c06-456e-b570-5b3774a834d9\") " pod="metallb-system/controller-7bb4cc7c98-47vwc" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.511495 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8p7dn\" (UniqueName: \"kubernetes.io/projected/a4d6dd0e-4763-486d-a6a6-8ce519149aaa-kube-api-access-8p7dn\") pod \"speaker-rrmgz\" (UID: \"a4d6dd0e-4763-486d-a6a6-8ce519149aaa\") " pod="metallb-system/speaker-rrmgz" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.511689 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlx6z\" (UniqueName: \"kubernetes.io/projected/475686c9-3c06-456e-b570-5b3774a834d9-kube-api-access-dlx6z\") pod \"controller-7bb4cc7c98-47vwc\" (UID: \"475686c9-3c06-456e-b570-5b3774a834d9\") " pod="metallb-system/controller-7bb4cc7c98-47vwc" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.602184 4923 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.620729 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-47vwc" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.797426 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0085e43b-b941-4b8b-b0c3-3ca1cdd68c58-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-zwkrg\" (UID: \"0085e43b-b941-4b8b-b0c3-3ca1cdd68c58\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-zwkrg" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.805264 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0085e43b-b941-4b8b-b0c3-3ca1cdd68c58-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-zwkrg\" (UID: \"0085e43b-b941-4b8b-b0c3-3ca1cdd68c58\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-zwkrg" Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.809977 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-47vwc"] Mar 11 13:35:18 crc kubenswrapper[4923]: W0311 13:35:18.817120 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod475686c9_3c06_456e_b570_5b3774a834d9.slice/crio-59fb7425174e416c7fb6ad87536ade132eef1711eb78f5966085a9b7b732370e WatchSource:0}: Error finding container 59fb7425174e416c7fb6ad87536ade132eef1711eb78f5966085a9b7b732370e: Status 404 returned error can't find the container with id 59fb7425174e416c7fb6ad87536ade132eef1711eb78f5966085a9b7b732370e Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.839790 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-47vwc" event={"ID":"475686c9-3c06-456e-b570-5b3774a834d9","Type":"ContainerStarted","Data":"59fb7425174e416c7fb6ad87536ade132eef1711eb78f5966085a9b7b732370e"} Mar 11 13:35:18 crc kubenswrapper[4923]: I0311 13:35:18.843045 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9f8w7" event={"ID":"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0","Type":"ContainerStarted","Data":"89152d7c687bdd6fd5867fd144b58c9d58eb801636075c696a7b11ebee84e430"} Mar 11 13:35:19 crc kubenswrapper[4923]: I0311 13:35:19.000056 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a4d6dd0e-4763-486d-a6a6-8ce519149aaa-memberlist\") pod \"speaker-rrmgz\" (UID: \"a4d6dd0e-4763-486d-a6a6-8ce519149aaa\") " pod="metallb-system/speaker-rrmgz" Mar 11 13:35:19 crc kubenswrapper[4923]: I0311 13:35:19.000123 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4d6dd0e-4763-486d-a6a6-8ce519149aaa-metrics-certs\") pod \"speaker-rrmgz\" (UID: \"a4d6dd0e-4763-486d-a6a6-8ce519149aaa\") " pod="metallb-system/speaker-rrmgz" Mar 11 13:35:19 crc kubenswrapper[4923]: E0311 13:35:19.000251 4923 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 11 13:35:19 crc kubenswrapper[4923]: E0311 13:35:19.000438 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a4d6dd0e-4763-486d-a6a6-8ce519149aaa-memberlist podName:a4d6dd0e-4763-486d-a6a6-8ce519149aaa nodeName:}" failed. No retries permitted until 2026-03-11 13:35:20.000407985 +0000 UTC m=+767.682309729 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/a4d6dd0e-4763-486d-a6a6-8ce519149aaa-memberlist") pod "speaker-rrmgz" (UID: "a4d6dd0e-4763-486d-a6a6-8ce519149aaa") : secret "metallb-memberlist" not found Mar 11 13:35:19 crc kubenswrapper[4923]: I0311 13:35:19.005122 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a4d6dd0e-4763-486d-a6a6-8ce519149aaa-metrics-certs\") pod \"speaker-rrmgz\" (UID: \"a4d6dd0e-4763-486d-a6a6-8ce519149aaa\") " pod="metallb-system/speaker-rrmgz" Mar 11 13:35:19 crc kubenswrapper[4923]: I0311 13:35:19.036298 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-zwkrg" Mar 11 13:35:19 crc kubenswrapper[4923]: I0311 13:35:19.450195 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-zwkrg"] Mar 11 13:35:19 crc kubenswrapper[4923]: W0311 13:35:19.456529 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0085e43b_b941_4b8b_b0c3_3ca1cdd68c58.slice/crio-969c8af13a5ec924cb470918d355c17313cfc9b5e410c0d332068f5fcc295517 WatchSource:0}: Error finding container 969c8af13a5ec924cb470918d355c17313cfc9b5e410c0d332068f5fcc295517: Status 404 returned error can't find the container with id 969c8af13a5ec924cb470918d355c17313cfc9b5e410c0d332068f5fcc295517 Mar 11 13:35:19 crc kubenswrapper[4923]: I0311 13:35:19.851101 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-zwkrg" event={"ID":"0085e43b-b941-4b8b-b0c3-3ca1cdd68c58","Type":"ContainerStarted","Data":"969c8af13a5ec924cb470918d355c17313cfc9b5e410c0d332068f5fcc295517"} Mar 11 13:35:19 crc kubenswrapper[4923]: I0311 13:35:19.852996 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-47vwc" event={"ID":"475686c9-3c06-456e-b570-5b3774a834d9","Type":"ContainerStarted","Data":"fecac30ab6de609eb11bea7ab5d978c58a8829e0f48fe06394a3ba3f247f210d"} Mar 11 13:35:20 crc kubenswrapper[4923]: I0311 13:35:20.012711 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a4d6dd0e-4763-486d-a6a6-8ce519149aaa-memberlist\") pod \"speaker-rrmgz\" (UID: \"a4d6dd0e-4763-486d-a6a6-8ce519149aaa\") " pod="metallb-system/speaker-rrmgz" Mar 11 13:35:20 crc kubenswrapper[4923]: I0311 13:35:20.032307 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a4d6dd0e-4763-486d-a6a6-8ce519149aaa-memberlist\") pod \"speaker-rrmgz\" (UID: \"a4d6dd0e-4763-486d-a6a6-8ce519149aaa\") " pod="metallb-system/speaker-rrmgz" Mar 11 13:35:20 crc kubenswrapper[4923]: I0311 13:35:20.081764 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-rrmgz" Mar 11 13:35:20 crc kubenswrapper[4923]: W0311 13:35:20.099546 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4d6dd0e_4763_486d_a6a6_8ce519149aaa.slice/crio-69941aca619b01fc2dcea3db8448e47c23ea9371187c40ffbf9a7199dafda271 WatchSource:0}: Error finding container 69941aca619b01fc2dcea3db8448e47c23ea9371187c40ffbf9a7199dafda271: Status 404 returned error can't find the container with id 69941aca619b01fc2dcea3db8448e47c23ea9371187c40ffbf9a7199dafda271 Mar 11 13:35:20 crc kubenswrapper[4923]: I0311 13:35:20.867564 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-rrmgz" event={"ID":"a4d6dd0e-4763-486d-a6a6-8ce519149aaa","Type":"ContainerStarted","Data":"86f4a181fbb82759d1205f4e41963a77505cd147e06920b826b9bf82f7729bc1"} Mar 11 13:35:20 crc kubenswrapper[4923]: I0311 13:35:20.867614 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-rrmgz" event={"ID":"a4d6dd0e-4763-486d-a6a6-8ce519149aaa","Type":"ContainerStarted","Data":"69941aca619b01fc2dcea3db8448e47c23ea9371187c40ffbf9a7199dafda271"} Mar 11 13:35:23 crc kubenswrapper[4923]: I0311 13:35:23.893864 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-47vwc" event={"ID":"475686c9-3c06-456e-b570-5b3774a834d9","Type":"ContainerStarted","Data":"eea29ac64e6eac7dd59b3f1edf4d46213a9d51d860e50f9729f4a6d9642241c7"} Mar 11 13:35:23 crc kubenswrapper[4923]: I0311 13:35:23.895270 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-7bb4cc7c98-47vwc" Mar 11 13:35:23 crc kubenswrapper[4923]: I0311 13:35:23.898325 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-rrmgz" event={"ID":"a4d6dd0e-4763-486d-a6a6-8ce519149aaa","Type":"ContainerStarted","Data":"affeec94272238de261f1dd0f1ceb8c818b1eb7b28ccaeba302f1d1a27876a33"} Mar 11 13:35:23 crc kubenswrapper[4923]: I0311 13:35:23.898482 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-rrmgz" Mar 11 13:35:23 crc kubenswrapper[4923]: I0311 13:35:23.911547 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-7bb4cc7c98-47vwc" podStartSLOduration=1.968569212 podStartE2EDuration="5.911526681s" podCreationTimestamp="2026-03-11 13:35:18 +0000 UTC" firstStartedPulling="2026-03-11 13:35:18.931648379 +0000 UTC m=+766.613550103" lastFinishedPulling="2026-03-11 13:35:22.874605858 +0000 UTC m=+770.556507572" observedRunningTime="2026-03-11 13:35:23.909389991 +0000 UTC m=+771.591291705" watchObservedRunningTime="2026-03-11 13:35:23.911526681 +0000 UTC m=+771.593428405" Mar 11 13:35:23 crc kubenswrapper[4923]: I0311 13:35:23.931331 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-rrmgz" podStartSLOduration=3.460002337 podStartE2EDuration="5.931316118s" podCreationTimestamp="2026-03-11 13:35:18 +0000 UTC" firstStartedPulling="2026-03-11 13:35:20.410437948 +0000 UTC m=+768.092339662" lastFinishedPulling="2026-03-11 13:35:22.881751739 +0000 UTC m=+770.563653443" observedRunningTime="2026-03-11 13:35:23.926762749 +0000 UTC m=+771.608664463" watchObservedRunningTime="2026-03-11 13:35:23.931316118 +0000 UTC m=+771.613217832" Mar 11 13:35:26 crc kubenswrapper[4923]: I0311 13:35:26.918961 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-zwkrg" event={"ID":"0085e43b-b941-4b8b-b0c3-3ca1cdd68c58","Type":"ContainerStarted","Data":"99a90531a96be8e1934d35a757a2072b9119cc6041daa742d7168b55b06f9050"} Mar 11 13:35:26 crc kubenswrapper[4923]: I0311 13:35:26.920932 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-zwkrg" Mar 11 13:35:26 crc kubenswrapper[4923]: I0311 13:35:26.921738 4923 generic.go:334] "Generic (PLEG): container finished" podID="21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0" containerID="b02150273189a2dd728b0dbedc85edd84f9b90525cf57ce067d17310851c6908" exitCode=0 Mar 11 13:35:26 crc kubenswrapper[4923]: I0311 13:35:26.921782 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9f8w7" event={"ID":"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0","Type":"ContainerDied","Data":"b02150273189a2dd728b0dbedc85edd84f9b90525cf57ce067d17310851c6908"} Mar 11 13:35:26 crc kubenswrapper[4923]: I0311 13:35:26.962450 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-zwkrg" podStartSLOduration=2.451327609 podStartE2EDuration="8.962425774s" podCreationTimestamp="2026-03-11 13:35:18 +0000 UTC" firstStartedPulling="2026-03-11 13:35:19.460664818 +0000 UTC m=+767.142566532" lastFinishedPulling="2026-03-11 13:35:25.971762953 +0000 UTC m=+773.653664697" observedRunningTime="2026-03-11 13:35:26.951737723 +0000 UTC m=+774.633639467" watchObservedRunningTime="2026-03-11 13:35:26.962425774 +0000 UTC m=+774.644327528" Mar 11 13:35:27 crc kubenswrapper[4923]: I0311 13:35:27.935379 4923 generic.go:334] "Generic (PLEG): container finished" podID="21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0" containerID="9e5bfe8192110394aada10dec8a3c9a6fa906351b3f2de9b688c1229913df0da" exitCode=0 Mar 11 13:35:27 crc kubenswrapper[4923]: I0311 13:35:27.936356 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9f8w7" event={"ID":"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0","Type":"ContainerDied","Data":"9e5bfe8192110394aada10dec8a3c9a6fa906351b3f2de9b688c1229913df0da"} Mar 11 13:35:28 crc kubenswrapper[4923]: I0311 13:35:28.627071 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-7bb4cc7c98-47vwc" Mar 11 13:35:28 crc kubenswrapper[4923]: I0311 13:35:28.945333 4923 generic.go:334] "Generic (PLEG): container finished" podID="21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0" containerID="7cfd6838185db24d5eb0f0d65711f7c6de5bf32746179bc303d4c93b864e4252" exitCode=0 Mar 11 13:35:28 crc kubenswrapper[4923]: I0311 13:35:28.945420 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9f8w7" event={"ID":"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0","Type":"ContainerDied","Data":"7cfd6838185db24d5eb0f0d65711f7c6de5bf32746179bc303d4c93b864e4252"} Mar 11 13:35:29 crc kubenswrapper[4923]: I0311 13:35:29.957155 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9f8w7" event={"ID":"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0","Type":"ContainerStarted","Data":"e64ff8cff52d4e6def7113836450c95663543cd95e6d2b83e32c6440fb063256"} Mar 11 13:35:29 crc kubenswrapper[4923]: I0311 13:35:29.957536 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9f8w7" event={"ID":"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0","Type":"ContainerStarted","Data":"22c967f53679a364ab9bb9cc7689f6894ff9733e0736212da17e37607de07e56"} Mar 11 13:35:29 crc kubenswrapper[4923]: I0311 13:35:29.957552 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9f8w7" event={"ID":"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0","Type":"ContainerStarted","Data":"496ae8f754f7a08a26d0475e92636aa7707c14dbfdae7695bd338fe886a2c695"} Mar 11 13:35:29 crc kubenswrapper[4923]: I0311 13:35:29.957562 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9f8w7" event={"ID":"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0","Type":"ContainerStarted","Data":"c553a15e9a8186c8b263c0d9b01ddadbc56e889398e8638e814bfe8230919d6e"} Mar 11 13:35:29 crc kubenswrapper[4923]: I0311 13:35:29.957570 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9f8w7" event={"ID":"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0","Type":"ContainerStarted","Data":"ff6579be31456e3c185992742b67c64bc2b1cef0680e7c334d27ed6cef79dc9d"} Mar 11 13:35:30 crc kubenswrapper[4923]: I0311 13:35:30.085694 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-rrmgz" Mar 11 13:35:30 crc kubenswrapper[4923]: I0311 13:35:30.972784 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-9f8w7" event={"ID":"21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0","Type":"ContainerStarted","Data":"a28b1cbd73c547a878db17db54adeb2be8fa138de602538ee67a903994c17782"} Mar 11 13:35:30 crc kubenswrapper[4923]: I0311 13:35:30.973562 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:31 crc kubenswrapper[4923]: I0311 13:35:31.006994 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-9f8w7" podStartSLOduration=5.592792001 podStartE2EDuration="13.006971952s" podCreationTimestamp="2026-03-11 13:35:18 +0000 UTC" firstStartedPulling="2026-03-11 13:35:18.602004982 +0000 UTC m=+766.283906696" lastFinishedPulling="2026-03-11 13:35:26.016184893 +0000 UTC m=+773.698086647" observedRunningTime="2026-03-11 13:35:31.00582798 +0000 UTC m=+778.687729734" watchObservedRunningTime="2026-03-11 13:35:31.006971952 +0000 UTC m=+778.688873696" Mar 11 13:35:33 crc kubenswrapper[4923]: I0311 13:35:33.458897 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:33 crc kubenswrapper[4923]: I0311 13:35:33.528842 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:35 crc kubenswrapper[4923]: I0311 13:35:35.969417 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-69cl4"] Mar 11 13:35:35 crc kubenswrapper[4923]: I0311 13:35:35.973261 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-69cl4" Mar 11 13:35:35 crc kubenswrapper[4923]: I0311 13:35:35.976006 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-wh5zq" Mar 11 13:35:35 crc kubenswrapper[4923]: I0311 13:35:35.978906 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 11 13:35:35 crc kubenswrapper[4923]: I0311 13:35:35.978969 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 11 13:35:36 crc kubenswrapper[4923]: I0311 13:35:36.001004 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-69cl4"] Mar 11 13:35:36 crc kubenswrapper[4923]: I0311 13:35:36.162300 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqt6c\" (UniqueName: \"kubernetes.io/projected/3ccd2679-f26d-4c0b-af0e-82f98d69fdd1-kube-api-access-zqt6c\") pod \"mariadb-operator-index-69cl4\" (UID: \"3ccd2679-f26d-4c0b-af0e-82f98d69fdd1\") " pod="openstack-operators/mariadb-operator-index-69cl4" Mar 11 13:35:36 crc kubenswrapper[4923]: I0311 13:35:36.263985 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqt6c\" (UniqueName: \"kubernetes.io/projected/3ccd2679-f26d-4c0b-af0e-82f98d69fdd1-kube-api-access-zqt6c\") pod \"mariadb-operator-index-69cl4\" (UID: \"3ccd2679-f26d-4c0b-af0e-82f98d69fdd1\") " pod="openstack-operators/mariadb-operator-index-69cl4" Mar 11 13:35:36 crc kubenswrapper[4923]: I0311 13:35:36.300417 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqt6c\" (UniqueName: \"kubernetes.io/projected/3ccd2679-f26d-4c0b-af0e-82f98d69fdd1-kube-api-access-zqt6c\") pod \"mariadb-operator-index-69cl4\" (UID: \"3ccd2679-f26d-4c0b-af0e-82f98d69fdd1\") " pod="openstack-operators/mariadb-operator-index-69cl4" Mar 11 13:35:36 crc kubenswrapper[4923]: I0311 13:35:36.306587 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-69cl4" Mar 11 13:35:36 crc kubenswrapper[4923]: I0311 13:35:36.777558 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-69cl4"] Mar 11 13:35:36 crc kubenswrapper[4923]: W0311 13:35:36.784127 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ccd2679_f26d_4c0b_af0e_82f98d69fdd1.slice/crio-edb81c4ddbe4b8aa1bd3ccc1d42d27e5a6b16f3bebda1e086ff96b6766d883cc WatchSource:0}: Error finding container edb81c4ddbe4b8aa1bd3ccc1d42d27e5a6b16f3bebda1e086ff96b6766d883cc: Status 404 returned error can't find the container with id edb81c4ddbe4b8aa1bd3ccc1d42d27e5a6b16f3bebda1e086ff96b6766d883cc Mar 11 13:35:37 crc kubenswrapper[4923]: I0311 13:35:37.017893 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-69cl4" event={"ID":"3ccd2679-f26d-4c0b-af0e-82f98d69fdd1","Type":"ContainerStarted","Data":"edb81c4ddbe4b8aa1bd3ccc1d42d27e5a6b16f3bebda1e086ff96b6766d883cc"} Mar 11 13:35:38 crc kubenswrapper[4923]: I0311 13:35:38.941973 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-69cl4"] Mar 11 13:35:39 crc kubenswrapper[4923]: I0311 13:35:39.042909 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-zwkrg" Mar 11 13:35:39 crc kubenswrapper[4923]: I0311 13:35:39.745537 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-btj29"] Mar 11 13:35:39 crc kubenswrapper[4923]: I0311 13:35:39.746567 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-btj29" Mar 11 13:35:39 crc kubenswrapper[4923]: I0311 13:35:39.766134 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-btj29"] Mar 11 13:35:39 crc kubenswrapper[4923]: I0311 13:35:39.916544 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7htf\" (UniqueName: \"kubernetes.io/projected/8f258a09-0cb8-43c1-9f1b-72b75147353c-kube-api-access-j7htf\") pod \"mariadb-operator-index-btj29\" (UID: \"8f258a09-0cb8-43c1-9f1b-72b75147353c\") " pod="openstack-operators/mariadb-operator-index-btj29" Mar 11 13:35:40 crc kubenswrapper[4923]: I0311 13:35:40.018046 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7htf\" (UniqueName: \"kubernetes.io/projected/8f258a09-0cb8-43c1-9f1b-72b75147353c-kube-api-access-j7htf\") pod \"mariadb-operator-index-btj29\" (UID: \"8f258a09-0cb8-43c1-9f1b-72b75147353c\") " pod="openstack-operators/mariadb-operator-index-btj29" Mar 11 13:35:40 crc kubenswrapper[4923]: I0311 13:35:40.040996 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-69cl4" event={"ID":"3ccd2679-f26d-4c0b-af0e-82f98d69fdd1","Type":"ContainerStarted","Data":"64a2f2caa9a67daa22f278016cf65257a3dd74a5b4498852afa223547af1f894"} Mar 11 13:35:40 crc kubenswrapper[4923]: I0311 13:35:40.041166 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-69cl4" podUID="3ccd2679-f26d-4c0b-af0e-82f98d69fdd1" containerName="registry-server" containerID="cri-o://64a2f2caa9a67daa22f278016cf65257a3dd74a5b4498852afa223547af1f894" gracePeriod=2 Mar 11 13:35:40 crc kubenswrapper[4923]: I0311 13:35:40.052214 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7htf\" (UniqueName: \"kubernetes.io/projected/8f258a09-0cb8-43c1-9f1b-72b75147353c-kube-api-access-j7htf\") pod \"mariadb-operator-index-btj29\" (UID: \"8f258a09-0cb8-43c1-9f1b-72b75147353c\") " pod="openstack-operators/mariadb-operator-index-btj29" Mar 11 13:35:40 crc kubenswrapper[4923]: I0311 13:35:40.067453 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-69cl4" podStartSLOduration=2.541714077 podStartE2EDuration="5.067417991s" podCreationTimestamp="2026-03-11 13:35:35 +0000 UTC" firstStartedPulling="2026-03-11 13:35:36.788219014 +0000 UTC m=+784.470120738" lastFinishedPulling="2026-03-11 13:35:39.313922938 +0000 UTC m=+786.995824652" observedRunningTime="2026-03-11 13:35:40.056933548 +0000 UTC m=+787.738835302" watchObservedRunningTime="2026-03-11 13:35:40.067417991 +0000 UTC m=+787.749319735" Mar 11 13:35:40 crc kubenswrapper[4923]: I0311 13:35:40.068182 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-btj29" Mar 11 13:35:40 crc kubenswrapper[4923]: I0311 13:35:40.249544 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-btj29"] Mar 11 13:35:40 crc kubenswrapper[4923]: W0311 13:35:40.260290 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f258a09_0cb8_43c1_9f1b_72b75147353c.slice/crio-94784c160083e9db0a9f4fd88121b667b6de1f42ac3570dec0f2d0b41b3abbae WatchSource:0}: Error finding container 94784c160083e9db0a9f4fd88121b667b6de1f42ac3570dec0f2d0b41b3abbae: Status 404 returned error can't find the container with id 94784c160083e9db0a9f4fd88121b667b6de1f42ac3570dec0f2d0b41b3abbae Mar 11 13:35:40 crc kubenswrapper[4923]: I0311 13:35:40.484385 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-69cl4" Mar 11 13:35:40 crc kubenswrapper[4923]: I0311 13:35:40.630796 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqt6c\" (UniqueName: \"kubernetes.io/projected/3ccd2679-f26d-4c0b-af0e-82f98d69fdd1-kube-api-access-zqt6c\") pod \"3ccd2679-f26d-4c0b-af0e-82f98d69fdd1\" (UID: \"3ccd2679-f26d-4c0b-af0e-82f98d69fdd1\") " Mar 11 13:35:40 crc kubenswrapper[4923]: I0311 13:35:40.638052 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ccd2679-f26d-4c0b-af0e-82f98d69fdd1-kube-api-access-zqt6c" (OuterVolumeSpecName: "kube-api-access-zqt6c") pod "3ccd2679-f26d-4c0b-af0e-82f98d69fdd1" (UID: "3ccd2679-f26d-4c0b-af0e-82f98d69fdd1"). InnerVolumeSpecName "kube-api-access-zqt6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:35:40 crc kubenswrapper[4923]: I0311 13:35:40.733273 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqt6c\" (UniqueName: \"kubernetes.io/projected/3ccd2679-f26d-4c0b-af0e-82f98d69fdd1-kube-api-access-zqt6c\") on node \"crc\" DevicePath \"\"" Mar 11 13:35:41 crc kubenswrapper[4923]: I0311 13:35:41.048310 4923 generic.go:334] "Generic (PLEG): container finished" podID="3ccd2679-f26d-4c0b-af0e-82f98d69fdd1" containerID="64a2f2caa9a67daa22f278016cf65257a3dd74a5b4498852afa223547af1f894" exitCode=0 Mar 11 13:35:41 crc kubenswrapper[4923]: I0311 13:35:41.048409 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-69cl4" event={"ID":"3ccd2679-f26d-4c0b-af0e-82f98d69fdd1","Type":"ContainerDied","Data":"64a2f2caa9a67daa22f278016cf65257a3dd74a5b4498852afa223547af1f894"} Mar 11 13:35:41 crc kubenswrapper[4923]: I0311 13:35:41.048444 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-69cl4" event={"ID":"3ccd2679-f26d-4c0b-af0e-82f98d69fdd1","Type":"ContainerDied","Data":"edb81c4ddbe4b8aa1bd3ccc1d42d27e5a6b16f3bebda1e086ff96b6766d883cc"} Mar 11 13:35:41 crc kubenswrapper[4923]: I0311 13:35:41.048465 4923 scope.go:117] "RemoveContainer" containerID="64a2f2caa9a67daa22f278016cf65257a3dd74a5b4498852afa223547af1f894" Mar 11 13:35:41 crc kubenswrapper[4923]: I0311 13:35:41.048465 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-69cl4" Mar 11 13:35:41 crc kubenswrapper[4923]: I0311 13:35:41.051190 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-btj29" event={"ID":"8f258a09-0cb8-43c1-9f1b-72b75147353c","Type":"ContainerStarted","Data":"94784c160083e9db0a9f4fd88121b667b6de1f42ac3570dec0f2d0b41b3abbae"} Mar 11 13:35:41 crc kubenswrapper[4923]: I0311 13:35:41.075613 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-69cl4"] Mar 11 13:35:41 crc kubenswrapper[4923]: I0311 13:35:41.079721 4923 scope.go:117] "RemoveContainer" containerID="64a2f2caa9a67daa22f278016cf65257a3dd74a5b4498852afa223547af1f894" Mar 11 13:35:41 crc kubenswrapper[4923]: I0311 13:35:41.080674 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-69cl4"] Mar 11 13:35:41 crc kubenswrapper[4923]: E0311 13:35:41.080710 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64a2f2caa9a67daa22f278016cf65257a3dd74a5b4498852afa223547af1f894\": container with ID starting with 64a2f2caa9a67daa22f278016cf65257a3dd74a5b4498852afa223547af1f894 not found: ID does not exist" containerID="64a2f2caa9a67daa22f278016cf65257a3dd74a5b4498852afa223547af1f894" Mar 11 13:35:41 crc kubenswrapper[4923]: I0311 13:35:41.080756 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64a2f2caa9a67daa22f278016cf65257a3dd74a5b4498852afa223547af1f894"} err="failed to get container status \"64a2f2caa9a67daa22f278016cf65257a3dd74a5b4498852afa223547af1f894\": rpc error: code = NotFound desc = could not find container \"64a2f2caa9a67daa22f278016cf65257a3dd74a5b4498852afa223547af1f894\": container with ID starting with 64a2f2caa9a67daa22f278016cf65257a3dd74a5b4498852afa223547af1f894 not found: ID does not exist" Mar 11 13:35:42 crc kubenswrapper[4923]: I0311 13:35:42.562822 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qzzpj"] Mar 11 13:35:42 crc kubenswrapper[4923]: E0311 13:35:42.564019 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ccd2679-f26d-4c0b-af0e-82f98d69fdd1" containerName="registry-server" Mar 11 13:35:42 crc kubenswrapper[4923]: I0311 13:35:42.564056 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ccd2679-f26d-4c0b-af0e-82f98d69fdd1" containerName="registry-server" Mar 11 13:35:42 crc kubenswrapper[4923]: I0311 13:35:42.564533 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ccd2679-f26d-4c0b-af0e-82f98d69fdd1" containerName="registry-server" Mar 11 13:35:42 crc kubenswrapper[4923]: I0311 13:35:42.568017 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzzpj" Mar 11 13:35:42 crc kubenswrapper[4923]: I0311 13:35:42.600675 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzzpj"] Mar 11 13:35:42 crc kubenswrapper[4923]: I0311 13:35:42.663145 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/131e0fb1-3ecc-4e82-8275-4c4f8ff434c2-utilities\") pod \"redhat-marketplace-qzzpj\" (UID: \"131e0fb1-3ecc-4e82-8275-4c4f8ff434c2\") " pod="openshift-marketplace/redhat-marketplace-qzzpj" Mar 11 13:35:42 crc kubenswrapper[4923]: I0311 13:35:42.663435 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/131e0fb1-3ecc-4e82-8275-4c4f8ff434c2-catalog-content\") pod \"redhat-marketplace-qzzpj\" (UID: \"131e0fb1-3ecc-4e82-8275-4c4f8ff434c2\") " pod="openshift-marketplace/redhat-marketplace-qzzpj" Mar 11 13:35:42 crc kubenswrapper[4923]: I0311 13:35:42.663601 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dz8f\" (UniqueName: \"kubernetes.io/projected/131e0fb1-3ecc-4e82-8275-4c4f8ff434c2-kube-api-access-5dz8f\") pod \"redhat-marketplace-qzzpj\" (UID: \"131e0fb1-3ecc-4e82-8275-4c4f8ff434c2\") " pod="openshift-marketplace/redhat-marketplace-qzzpj" Mar 11 13:35:42 crc kubenswrapper[4923]: I0311 13:35:42.765244 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/131e0fb1-3ecc-4e82-8275-4c4f8ff434c2-catalog-content\") pod \"redhat-marketplace-qzzpj\" (UID: \"131e0fb1-3ecc-4e82-8275-4c4f8ff434c2\") " pod="openshift-marketplace/redhat-marketplace-qzzpj" Mar 11 13:35:42 crc kubenswrapper[4923]: I0311 13:35:42.765663 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dz8f\" (UniqueName: \"kubernetes.io/projected/131e0fb1-3ecc-4e82-8275-4c4f8ff434c2-kube-api-access-5dz8f\") pod \"redhat-marketplace-qzzpj\" (UID: \"131e0fb1-3ecc-4e82-8275-4c4f8ff434c2\") " pod="openshift-marketplace/redhat-marketplace-qzzpj" Mar 11 13:35:42 crc kubenswrapper[4923]: I0311 13:35:42.765724 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/131e0fb1-3ecc-4e82-8275-4c4f8ff434c2-utilities\") pod \"redhat-marketplace-qzzpj\" (UID: \"131e0fb1-3ecc-4e82-8275-4c4f8ff434c2\") " pod="openshift-marketplace/redhat-marketplace-qzzpj" Mar 11 13:35:42 crc kubenswrapper[4923]: I0311 13:35:42.766530 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/131e0fb1-3ecc-4e82-8275-4c4f8ff434c2-utilities\") pod \"redhat-marketplace-qzzpj\" (UID: \"131e0fb1-3ecc-4e82-8275-4c4f8ff434c2\") " pod="openshift-marketplace/redhat-marketplace-qzzpj" Mar 11 13:35:42 crc kubenswrapper[4923]: I0311 13:35:42.766537 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/131e0fb1-3ecc-4e82-8275-4c4f8ff434c2-catalog-content\") pod \"redhat-marketplace-qzzpj\" (UID: \"131e0fb1-3ecc-4e82-8275-4c4f8ff434c2\") " pod="openshift-marketplace/redhat-marketplace-qzzpj" Mar 11 13:35:42 crc kubenswrapper[4923]: I0311 13:35:42.799039 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dz8f\" (UniqueName: \"kubernetes.io/projected/131e0fb1-3ecc-4e82-8275-4c4f8ff434c2-kube-api-access-5dz8f\") pod \"redhat-marketplace-qzzpj\" (UID: \"131e0fb1-3ecc-4e82-8275-4c4f8ff434c2\") " pod="openshift-marketplace/redhat-marketplace-qzzpj" Mar 11 13:35:42 crc kubenswrapper[4923]: I0311 13:35:42.935483 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzzpj" Mar 11 13:35:43 crc kubenswrapper[4923]: I0311 13:35:43.035283 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ccd2679-f26d-4c0b-af0e-82f98d69fdd1" path="/var/lib/kubelet/pods/3ccd2679-f26d-4c0b-af0e-82f98d69fdd1/volumes" Mar 11 13:35:43 crc kubenswrapper[4923]: I0311 13:35:43.071205 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-btj29" event={"ID":"8f258a09-0cb8-43c1-9f1b-72b75147353c","Type":"ContainerStarted","Data":"56cc833e2dea4d6d48e3e4ae7fe06873547a359445da82b1a2498dc937e81c94"} Mar 11 13:35:43 crc kubenswrapper[4923]: I0311 13:35:43.095575 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-btj29" podStartSLOduration=2.192114747 podStartE2EDuration="4.09554981s" podCreationTimestamp="2026-03-11 13:35:39 +0000 UTC" firstStartedPulling="2026-03-11 13:35:40.273581136 +0000 UTC m=+787.955482900" lastFinishedPulling="2026-03-11 13:35:42.177016209 +0000 UTC m=+789.858917963" observedRunningTime="2026-03-11 13:35:43.085500509 +0000 UTC m=+790.767402233" watchObservedRunningTime="2026-03-11 13:35:43.09554981 +0000 UTC m=+790.777451524" Mar 11 13:35:43 crc kubenswrapper[4923]: I0311 13:35:43.370695 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzzpj"] Mar 11 13:35:43 crc kubenswrapper[4923]: W0311 13:35:43.376766 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod131e0fb1_3ecc_4e82_8275_4c4f8ff434c2.slice/crio-10a0a7fd7e6766d7ae9713b672c9bbd9776883a5f9de30f5591093c02fcf3f29 WatchSource:0}: Error finding container 10a0a7fd7e6766d7ae9713b672c9bbd9776883a5f9de30f5591093c02fcf3f29: Status 404 returned error can't find the container with id 10a0a7fd7e6766d7ae9713b672c9bbd9776883a5f9de30f5591093c02fcf3f29 Mar 11 13:35:44 crc kubenswrapper[4923]: I0311 13:35:44.082806 4923 generic.go:334] "Generic (PLEG): container finished" podID="131e0fb1-3ecc-4e82-8275-4c4f8ff434c2" containerID="5d9ae1eac6aa6e1223fcc53a69e194ea6680c502c9f2f6ff06901fc67d42e8d5" exitCode=0 Mar 11 13:35:44 crc kubenswrapper[4923]: I0311 13:35:44.083119 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzzpj" event={"ID":"131e0fb1-3ecc-4e82-8275-4c4f8ff434c2","Type":"ContainerDied","Data":"5d9ae1eac6aa6e1223fcc53a69e194ea6680c502c9f2f6ff06901fc67d42e8d5"} Mar 11 13:35:44 crc kubenswrapper[4923]: I0311 13:35:44.083187 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzzpj" event={"ID":"131e0fb1-3ecc-4e82-8275-4c4f8ff434c2","Type":"ContainerStarted","Data":"10a0a7fd7e6766d7ae9713b672c9bbd9776883a5f9de30f5591093c02fcf3f29"} Mar 11 13:35:45 crc kubenswrapper[4923]: I0311 13:35:45.094515 4923 generic.go:334] "Generic (PLEG): container finished" podID="131e0fb1-3ecc-4e82-8275-4c4f8ff434c2" containerID="868744eb9aad075f3b23202e2667b8d98f45bd4c049f66cdefff4a420bfdf80b" exitCode=0 Mar 11 13:35:45 crc kubenswrapper[4923]: I0311 13:35:45.094676 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzzpj" event={"ID":"131e0fb1-3ecc-4e82-8275-4c4f8ff434c2","Type":"ContainerDied","Data":"868744eb9aad075f3b23202e2667b8d98f45bd4c049f66cdefff4a420bfdf80b"} Mar 11 13:35:46 crc kubenswrapper[4923]: I0311 13:35:46.103499 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzzpj" event={"ID":"131e0fb1-3ecc-4e82-8275-4c4f8ff434c2","Type":"ContainerStarted","Data":"efcca64b7c82afa6b44ecc3cf306b5ea5d53a91d84f018b38b8796ae0596ca2f"} Mar 11 13:35:46 crc kubenswrapper[4923]: I0311 13:35:46.127894 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qzzpj" podStartSLOduration=2.72090426 podStartE2EDuration="4.127872827s" podCreationTimestamp="2026-03-11 13:35:42 +0000 UTC" firstStartedPulling="2026-03-11 13:35:44.085338709 +0000 UTC m=+791.767240463" lastFinishedPulling="2026-03-11 13:35:45.492307286 +0000 UTC m=+793.174209030" observedRunningTime="2026-03-11 13:35:46.120964354 +0000 UTC m=+793.802866078" watchObservedRunningTime="2026-03-11 13:35:46.127872827 +0000 UTC m=+793.809774541" Mar 11 13:35:48 crc kubenswrapper[4923]: I0311 13:35:48.461970 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-9f8w7" Mar 11 13:35:50 crc kubenswrapper[4923]: I0311 13:35:50.069141 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-btj29" Mar 11 13:35:50 crc kubenswrapper[4923]: I0311 13:35:50.069686 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-btj29" Mar 11 13:35:50 crc kubenswrapper[4923]: I0311 13:35:50.117961 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-btj29" Mar 11 13:35:50 crc kubenswrapper[4923]: I0311 13:35:50.174985 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-btj29" Mar 11 13:35:51 crc kubenswrapper[4923]: I0311 13:35:51.986077 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2"] Mar 11 13:35:51 crc kubenswrapper[4923]: I0311 13:35:51.987725 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2" Mar 11 13:35:51 crc kubenswrapper[4923]: I0311 13:35:51.989997 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-qkqtx" Mar 11 13:35:51 crc kubenswrapper[4923]: I0311 13:35:51.998413 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2"] Mar 11 13:35:52 crc kubenswrapper[4923]: I0311 13:35:52.094711 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2726a877-fccf-4cd2-95a7-084920be7c21-bundle\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2\" (UID: \"2726a877-fccf-4cd2-95a7-084920be7c21\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2" Mar 11 13:35:52 crc kubenswrapper[4923]: I0311 13:35:52.094770 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2726a877-fccf-4cd2-95a7-084920be7c21-util\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2\" (UID: \"2726a877-fccf-4cd2-95a7-084920be7c21\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2" Mar 11 13:35:52 crc kubenswrapper[4923]: I0311 13:35:52.094812 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff7vg\" (UniqueName: \"kubernetes.io/projected/2726a877-fccf-4cd2-95a7-084920be7c21-kube-api-access-ff7vg\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2\" (UID: \"2726a877-fccf-4cd2-95a7-084920be7c21\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2" Mar 11 13:35:52 crc kubenswrapper[4923]: I0311 13:35:52.196604 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2726a877-fccf-4cd2-95a7-084920be7c21-util\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2\" (UID: \"2726a877-fccf-4cd2-95a7-084920be7c21\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2" Mar 11 13:35:52 crc kubenswrapper[4923]: I0311 13:35:52.196747 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff7vg\" (UniqueName: \"kubernetes.io/projected/2726a877-fccf-4cd2-95a7-084920be7c21-kube-api-access-ff7vg\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2\" (UID: \"2726a877-fccf-4cd2-95a7-084920be7c21\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2" Mar 11 13:35:52 crc kubenswrapper[4923]: I0311 13:35:52.196871 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2726a877-fccf-4cd2-95a7-084920be7c21-bundle\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2\" (UID: \"2726a877-fccf-4cd2-95a7-084920be7c21\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2" Mar 11 13:35:52 crc kubenswrapper[4923]: I0311 13:35:52.197586 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2726a877-fccf-4cd2-95a7-084920be7c21-util\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2\" (UID: \"2726a877-fccf-4cd2-95a7-084920be7c21\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2" Mar 11 13:35:52 crc kubenswrapper[4923]: I0311 13:35:52.197595 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2726a877-fccf-4cd2-95a7-084920be7c21-bundle\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2\" (UID: \"2726a877-fccf-4cd2-95a7-084920be7c21\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2" Mar 11 13:35:52 crc kubenswrapper[4923]: I0311 13:35:52.220156 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff7vg\" (UniqueName: \"kubernetes.io/projected/2726a877-fccf-4cd2-95a7-084920be7c21-kube-api-access-ff7vg\") pod \"b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2\" (UID: \"2726a877-fccf-4cd2-95a7-084920be7c21\") " pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2" Mar 11 13:35:52 crc kubenswrapper[4923]: I0311 13:35:52.309809 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2" Mar 11 13:35:52 crc kubenswrapper[4923]: I0311 13:35:52.745807 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2"] Mar 11 13:35:52 crc kubenswrapper[4923]: I0311 13:35:52.936497 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qzzpj" Mar 11 13:35:52 crc kubenswrapper[4923]: I0311 13:35:52.936565 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qzzpj" Mar 11 13:35:52 crc kubenswrapper[4923]: I0311 13:35:52.980951 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qzzpj" Mar 11 13:35:53 crc kubenswrapper[4923]: I0311 13:35:53.153211 4923 generic.go:334] "Generic (PLEG): container finished" podID="2726a877-fccf-4cd2-95a7-084920be7c21" containerID="816ee7aa415ef9e1817163896d4c983131fe11d0aa69994ab03d8983fca26370" exitCode=0 Mar 11 13:35:53 crc kubenswrapper[4923]: I0311 13:35:53.153502 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2" event={"ID":"2726a877-fccf-4cd2-95a7-084920be7c21","Type":"ContainerDied","Data":"816ee7aa415ef9e1817163896d4c983131fe11d0aa69994ab03d8983fca26370"} Mar 11 13:35:53 crc kubenswrapper[4923]: I0311 13:35:53.153599 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2" event={"ID":"2726a877-fccf-4cd2-95a7-084920be7c21","Type":"ContainerStarted","Data":"5a127376c8957ed35879ad5c7c2643b6fce13a35ddb3b441987b89166c1f7865"} Mar 11 13:35:53 crc kubenswrapper[4923]: I0311 13:35:53.252866 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qzzpj" Mar 11 13:35:54 crc kubenswrapper[4923]: I0311 13:35:54.164711 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2" event={"ID":"2726a877-fccf-4cd2-95a7-084920be7c21","Type":"ContainerStarted","Data":"fc5debf7322cc32fa341dec97e848a7031c93556c13bd1125e246396e854fb65"} Mar 11 13:35:55 crc kubenswrapper[4923]: I0311 13:35:55.179845 4923 generic.go:334] "Generic (PLEG): container finished" podID="2726a877-fccf-4cd2-95a7-084920be7c21" containerID="fc5debf7322cc32fa341dec97e848a7031c93556c13bd1125e246396e854fb65" exitCode=0 Mar 11 13:35:55 crc kubenswrapper[4923]: I0311 13:35:55.180172 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2" event={"ID":"2726a877-fccf-4cd2-95a7-084920be7c21","Type":"ContainerDied","Data":"fc5debf7322cc32fa341dec97e848a7031c93556c13bd1125e246396e854fb65"} Mar 11 13:35:56 crc kubenswrapper[4923]: I0311 13:35:56.192672 4923 generic.go:334] "Generic (PLEG): container finished" podID="2726a877-fccf-4cd2-95a7-084920be7c21" containerID="6b8da8c8c77849f749bb107daad66f8a59390e9e49e8782f7f8e15820593d973" exitCode=0 Mar 11 13:35:56 crc kubenswrapper[4923]: I0311 13:35:56.192752 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2" event={"ID":"2726a877-fccf-4cd2-95a7-084920be7c21","Type":"ContainerDied","Data":"6b8da8c8c77849f749bb107daad66f8a59390e9e49e8782f7f8e15820593d973"} Mar 11 13:35:56 crc kubenswrapper[4923]: I0311 13:35:56.540419 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzzpj"] Mar 11 13:35:56 crc kubenswrapper[4923]: I0311 13:35:56.540878 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qzzpj" podUID="131e0fb1-3ecc-4e82-8275-4c4f8ff434c2" containerName="registry-server" containerID="cri-o://efcca64b7c82afa6b44ecc3cf306b5ea5d53a91d84f018b38b8796ae0596ca2f" gracePeriod=2 Mar 11 13:35:56 crc kubenswrapper[4923]: I0311 13:35:56.964627 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzzpj" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.064962 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dz8f\" (UniqueName: \"kubernetes.io/projected/131e0fb1-3ecc-4e82-8275-4c4f8ff434c2-kube-api-access-5dz8f\") pod \"131e0fb1-3ecc-4e82-8275-4c4f8ff434c2\" (UID: \"131e0fb1-3ecc-4e82-8275-4c4f8ff434c2\") " Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.065043 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/131e0fb1-3ecc-4e82-8275-4c4f8ff434c2-utilities\") pod \"131e0fb1-3ecc-4e82-8275-4c4f8ff434c2\" (UID: \"131e0fb1-3ecc-4e82-8275-4c4f8ff434c2\") " Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.065151 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/131e0fb1-3ecc-4e82-8275-4c4f8ff434c2-catalog-content\") pod \"131e0fb1-3ecc-4e82-8275-4c4f8ff434c2\" (UID: \"131e0fb1-3ecc-4e82-8275-4c4f8ff434c2\") " Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.066521 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/131e0fb1-3ecc-4e82-8275-4c4f8ff434c2-utilities" (OuterVolumeSpecName: "utilities") pod "131e0fb1-3ecc-4e82-8275-4c4f8ff434c2" (UID: "131e0fb1-3ecc-4e82-8275-4c4f8ff434c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.070620 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/131e0fb1-3ecc-4e82-8275-4c4f8ff434c2-kube-api-access-5dz8f" (OuterVolumeSpecName: "kube-api-access-5dz8f") pod "131e0fb1-3ecc-4e82-8275-4c4f8ff434c2" (UID: "131e0fb1-3ecc-4e82-8275-4c4f8ff434c2"). InnerVolumeSpecName "kube-api-access-5dz8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.089276 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/131e0fb1-3ecc-4e82-8275-4c4f8ff434c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "131e0fb1-3ecc-4e82-8275-4c4f8ff434c2" (UID: "131e0fb1-3ecc-4e82-8275-4c4f8ff434c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.166759 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/131e0fb1-3ecc-4e82-8275-4c4f8ff434c2-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.166792 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dz8f\" (UniqueName: \"kubernetes.io/projected/131e0fb1-3ecc-4e82-8275-4c4f8ff434c2-kube-api-access-5dz8f\") on node \"crc\" DevicePath \"\"" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.166807 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/131e0fb1-3ecc-4e82-8275-4c4f8ff434c2-utilities\") on node \"crc\" DevicePath \"\"" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.201414 4923 generic.go:334] "Generic (PLEG): container finished" podID="131e0fb1-3ecc-4e82-8275-4c4f8ff434c2" containerID="efcca64b7c82afa6b44ecc3cf306b5ea5d53a91d84f018b38b8796ae0596ca2f" exitCode=0 Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.201518 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzzpj" event={"ID":"131e0fb1-3ecc-4e82-8275-4c4f8ff434c2","Type":"ContainerDied","Data":"efcca64b7c82afa6b44ecc3cf306b5ea5d53a91d84f018b38b8796ae0596ca2f"} Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.201571 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qzzpj" event={"ID":"131e0fb1-3ecc-4e82-8275-4c4f8ff434c2","Type":"ContainerDied","Data":"10a0a7fd7e6766d7ae9713b672c9bbd9776883a5f9de30f5591093c02fcf3f29"} Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.201529 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qzzpj" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.201593 4923 scope.go:117] "RemoveContainer" containerID="efcca64b7c82afa6b44ecc3cf306b5ea5d53a91d84f018b38b8796ae0596ca2f" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.228324 4923 scope.go:117] "RemoveContainer" containerID="868744eb9aad075f3b23202e2667b8d98f45bd4c049f66cdefff4a420bfdf80b" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.246415 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzzpj"] Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.257214 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qzzpj"] Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.285690 4923 scope.go:117] "RemoveContainer" containerID="5d9ae1eac6aa6e1223fcc53a69e194ea6680c502c9f2f6ff06901fc67d42e8d5" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.309115 4923 scope.go:117] "RemoveContainer" containerID="efcca64b7c82afa6b44ecc3cf306b5ea5d53a91d84f018b38b8796ae0596ca2f" Mar 11 13:35:57 crc kubenswrapper[4923]: E0311 13:35:57.311508 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"efcca64b7c82afa6b44ecc3cf306b5ea5d53a91d84f018b38b8796ae0596ca2f\": container with ID starting with efcca64b7c82afa6b44ecc3cf306b5ea5d53a91d84f018b38b8796ae0596ca2f not found: ID does not exist" containerID="efcca64b7c82afa6b44ecc3cf306b5ea5d53a91d84f018b38b8796ae0596ca2f" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.311553 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"efcca64b7c82afa6b44ecc3cf306b5ea5d53a91d84f018b38b8796ae0596ca2f"} err="failed to get container status \"efcca64b7c82afa6b44ecc3cf306b5ea5d53a91d84f018b38b8796ae0596ca2f\": rpc error: code = NotFound desc = could not find container \"efcca64b7c82afa6b44ecc3cf306b5ea5d53a91d84f018b38b8796ae0596ca2f\": container with ID starting with efcca64b7c82afa6b44ecc3cf306b5ea5d53a91d84f018b38b8796ae0596ca2f not found: ID does not exist" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.311578 4923 scope.go:117] "RemoveContainer" containerID="868744eb9aad075f3b23202e2667b8d98f45bd4c049f66cdefff4a420bfdf80b" Mar 11 13:35:57 crc kubenswrapper[4923]: E0311 13:35:57.312315 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"868744eb9aad075f3b23202e2667b8d98f45bd4c049f66cdefff4a420bfdf80b\": container with ID starting with 868744eb9aad075f3b23202e2667b8d98f45bd4c049f66cdefff4a420bfdf80b not found: ID does not exist" containerID="868744eb9aad075f3b23202e2667b8d98f45bd4c049f66cdefff4a420bfdf80b" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.312354 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"868744eb9aad075f3b23202e2667b8d98f45bd4c049f66cdefff4a420bfdf80b"} err="failed to get container status \"868744eb9aad075f3b23202e2667b8d98f45bd4c049f66cdefff4a420bfdf80b\": rpc error: code = NotFound desc = could not find container \"868744eb9aad075f3b23202e2667b8d98f45bd4c049f66cdefff4a420bfdf80b\": container with ID starting with 868744eb9aad075f3b23202e2667b8d98f45bd4c049f66cdefff4a420bfdf80b not found: ID does not exist" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.312373 4923 scope.go:117] "RemoveContainer" containerID="5d9ae1eac6aa6e1223fcc53a69e194ea6680c502c9f2f6ff06901fc67d42e8d5" Mar 11 13:35:57 crc kubenswrapper[4923]: E0311 13:35:57.312660 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d9ae1eac6aa6e1223fcc53a69e194ea6680c502c9f2f6ff06901fc67d42e8d5\": container with ID starting with 5d9ae1eac6aa6e1223fcc53a69e194ea6680c502c9f2f6ff06901fc67d42e8d5 not found: ID does not exist" containerID="5d9ae1eac6aa6e1223fcc53a69e194ea6680c502c9f2f6ff06901fc67d42e8d5" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.312679 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d9ae1eac6aa6e1223fcc53a69e194ea6680c502c9f2f6ff06901fc67d42e8d5"} err="failed to get container status \"5d9ae1eac6aa6e1223fcc53a69e194ea6680c502c9f2f6ff06901fc67d42e8d5\": rpc error: code = NotFound desc = could not find container \"5d9ae1eac6aa6e1223fcc53a69e194ea6680c502c9f2f6ff06901fc67d42e8d5\": container with ID starting with 5d9ae1eac6aa6e1223fcc53a69e194ea6680c502c9f2f6ff06901fc67d42e8d5 not found: ID does not exist" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.433875 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.570082 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2726a877-fccf-4cd2-95a7-084920be7c21-util\") pod \"2726a877-fccf-4cd2-95a7-084920be7c21\" (UID: \"2726a877-fccf-4cd2-95a7-084920be7c21\") " Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.570172 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2726a877-fccf-4cd2-95a7-084920be7c21-bundle\") pod \"2726a877-fccf-4cd2-95a7-084920be7c21\" (UID: \"2726a877-fccf-4cd2-95a7-084920be7c21\") " Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.570240 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ff7vg\" (UniqueName: \"kubernetes.io/projected/2726a877-fccf-4cd2-95a7-084920be7c21-kube-api-access-ff7vg\") pod \"2726a877-fccf-4cd2-95a7-084920be7c21\" (UID: \"2726a877-fccf-4cd2-95a7-084920be7c21\") " Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.571125 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2726a877-fccf-4cd2-95a7-084920be7c21-bundle" (OuterVolumeSpecName: "bundle") pod "2726a877-fccf-4cd2-95a7-084920be7c21" (UID: "2726a877-fccf-4cd2-95a7-084920be7c21"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.573258 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2726a877-fccf-4cd2-95a7-084920be7c21-kube-api-access-ff7vg" (OuterVolumeSpecName: "kube-api-access-ff7vg") pod "2726a877-fccf-4cd2-95a7-084920be7c21" (UID: "2726a877-fccf-4cd2-95a7-084920be7c21"). InnerVolumeSpecName "kube-api-access-ff7vg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.599977 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2726a877-fccf-4cd2-95a7-084920be7c21-util" (OuterVolumeSpecName: "util") pod "2726a877-fccf-4cd2-95a7-084920be7c21" (UID: "2726a877-fccf-4cd2-95a7-084920be7c21"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.671780 4923 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2726a877-fccf-4cd2-95a7-084920be7c21-util\") on node \"crc\" DevicePath \"\"" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.671828 4923 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2726a877-fccf-4cd2-95a7-084920be7c21-bundle\") on node \"crc\" DevicePath \"\"" Mar 11 13:35:57 crc kubenswrapper[4923]: I0311 13:35:57.671847 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ff7vg\" (UniqueName: \"kubernetes.io/projected/2726a877-fccf-4cd2-95a7-084920be7c21-kube-api-access-ff7vg\") on node \"crc\" DevicePath \"\"" Mar 11 13:35:58 crc kubenswrapper[4923]: I0311 13:35:58.211803 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2" event={"ID":"2726a877-fccf-4cd2-95a7-084920be7c21","Type":"ContainerDied","Data":"5a127376c8957ed35879ad5c7c2643b6fce13a35ddb3b441987b89166c1f7865"} Mar 11 13:35:58 crc kubenswrapper[4923]: I0311 13:35:58.211860 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a127376c8957ed35879ad5c7c2643b6fce13a35ddb3b441987b89166c1f7865" Mar 11 13:35:58 crc kubenswrapper[4923]: I0311 13:35:58.211909 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2" Mar 11 13:35:59 crc kubenswrapper[4923]: I0311 13:35:59.027667 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="131e0fb1-3ecc-4e82-8275-4c4f8ff434c2" path="/var/lib/kubelet/pods/131e0fb1-3ecc-4e82-8275-4c4f8ff434c2/volumes" Mar 11 13:36:00 crc kubenswrapper[4923]: I0311 13:36:00.148318 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29553936-k7s5k"] Mar 11 13:36:00 crc kubenswrapper[4923]: E0311 13:36:00.149232 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2726a877-fccf-4cd2-95a7-084920be7c21" containerName="extract" Mar 11 13:36:00 crc kubenswrapper[4923]: I0311 13:36:00.149259 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="2726a877-fccf-4cd2-95a7-084920be7c21" containerName="extract" Mar 11 13:36:00 crc kubenswrapper[4923]: E0311 13:36:00.149287 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="131e0fb1-3ecc-4e82-8275-4c4f8ff434c2" containerName="extract-utilities" Mar 11 13:36:00 crc kubenswrapper[4923]: I0311 13:36:00.149303 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="131e0fb1-3ecc-4e82-8275-4c4f8ff434c2" containerName="extract-utilities" Mar 11 13:36:00 crc kubenswrapper[4923]: E0311 13:36:00.149329 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2726a877-fccf-4cd2-95a7-084920be7c21" containerName="util" Mar 11 13:36:00 crc kubenswrapper[4923]: I0311 13:36:00.149377 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="2726a877-fccf-4cd2-95a7-084920be7c21" containerName="util" Mar 11 13:36:00 crc kubenswrapper[4923]: E0311 13:36:00.149409 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="131e0fb1-3ecc-4e82-8275-4c4f8ff434c2" containerName="extract-content" Mar 11 13:36:00 crc kubenswrapper[4923]: I0311 13:36:00.149424 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="131e0fb1-3ecc-4e82-8275-4c4f8ff434c2" containerName="extract-content" Mar 11 13:36:00 crc kubenswrapper[4923]: E0311 13:36:00.149443 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="131e0fb1-3ecc-4e82-8275-4c4f8ff434c2" containerName="registry-server" Mar 11 13:36:00 crc kubenswrapper[4923]: I0311 13:36:00.149462 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="131e0fb1-3ecc-4e82-8275-4c4f8ff434c2" containerName="registry-server" Mar 11 13:36:00 crc kubenswrapper[4923]: E0311 13:36:00.149492 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2726a877-fccf-4cd2-95a7-084920be7c21" containerName="pull" Mar 11 13:36:00 crc kubenswrapper[4923]: I0311 13:36:00.149507 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="2726a877-fccf-4cd2-95a7-084920be7c21" containerName="pull" Mar 11 13:36:00 crc kubenswrapper[4923]: I0311 13:36:00.149727 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="131e0fb1-3ecc-4e82-8275-4c4f8ff434c2" containerName="registry-server" Mar 11 13:36:00 crc kubenswrapper[4923]: I0311 13:36:00.149752 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="2726a877-fccf-4cd2-95a7-084920be7c21" containerName="extract" Mar 11 13:36:00 crc kubenswrapper[4923]: I0311 13:36:00.151386 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553936-k7s5k" Mar 11 13:36:00 crc kubenswrapper[4923]: I0311 13:36:00.154192 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-nxck7" Mar 11 13:36:00 crc kubenswrapper[4923]: I0311 13:36:00.154213 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 11 13:36:00 crc kubenswrapper[4923]: I0311 13:36:00.154246 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 11 13:36:00 crc kubenswrapper[4923]: I0311 13:36:00.160970 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553936-k7s5k"] Mar 11 13:36:00 crc kubenswrapper[4923]: I0311 13:36:00.311482 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dh4d\" (UniqueName: \"kubernetes.io/projected/85eb3ab9-5384-4c51-b20c-66eaa4f44c38-kube-api-access-4dh4d\") pod \"auto-csr-approver-29553936-k7s5k\" (UID: \"85eb3ab9-5384-4c51-b20c-66eaa4f44c38\") " pod="openshift-infra/auto-csr-approver-29553936-k7s5k" Mar 11 13:36:00 crc kubenswrapper[4923]: I0311 13:36:00.413080 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dh4d\" (UniqueName: \"kubernetes.io/projected/85eb3ab9-5384-4c51-b20c-66eaa4f44c38-kube-api-access-4dh4d\") pod \"auto-csr-approver-29553936-k7s5k\" (UID: \"85eb3ab9-5384-4c51-b20c-66eaa4f44c38\") " pod="openshift-infra/auto-csr-approver-29553936-k7s5k" Mar 11 13:36:00 crc kubenswrapper[4923]: I0311 13:36:00.440234 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dh4d\" (UniqueName: \"kubernetes.io/projected/85eb3ab9-5384-4c51-b20c-66eaa4f44c38-kube-api-access-4dh4d\") pod \"auto-csr-approver-29553936-k7s5k\" (UID: \"85eb3ab9-5384-4c51-b20c-66eaa4f44c38\") " pod="openshift-infra/auto-csr-approver-29553936-k7s5k" Mar 11 13:36:00 crc kubenswrapper[4923]: I0311 13:36:00.477553 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553936-k7s5k" Mar 11 13:36:00 crc kubenswrapper[4923]: I0311 13:36:00.928500 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553936-k7s5k"] Mar 11 13:36:01 crc kubenswrapper[4923]: I0311 13:36:01.231645 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553936-k7s5k" event={"ID":"85eb3ab9-5384-4c51-b20c-66eaa4f44c38","Type":"ContainerStarted","Data":"117ac95bd0af5823a2239dfecff35274d6288cb3cb25c4934fff31d4158967d8"} Mar 11 13:36:04 crc kubenswrapper[4923]: I0311 13:36:04.136988 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6d9597b6cb-f8lmr"] Mar 11 13:36:04 crc kubenswrapper[4923]: I0311 13:36:04.138139 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6d9597b6cb-f8lmr" Mar 11 13:36:04 crc kubenswrapper[4923]: I0311 13:36:04.139832 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Mar 11 13:36:04 crc kubenswrapper[4923]: I0311 13:36:04.140228 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Mar 11 13:36:04 crc kubenswrapper[4923]: I0311 13:36:04.140537 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-7cl7j" Mar 11 13:36:04 crc kubenswrapper[4923]: I0311 13:36:04.150024 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6d9597b6cb-f8lmr"] Mar 11 13:36:04 crc kubenswrapper[4923]: I0311 13:36:04.248690 4923 generic.go:334] "Generic (PLEG): container finished" podID="85eb3ab9-5384-4c51-b20c-66eaa4f44c38" containerID="9d946a60e806e217fa5f56b303307f58b580b8aff6f2cc08c8f6de3ffd1582a3" exitCode=0 Mar 11 13:36:04 crc kubenswrapper[4923]: I0311 13:36:04.248780 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553936-k7s5k" event={"ID":"85eb3ab9-5384-4c51-b20c-66eaa4f44c38","Type":"ContainerDied","Data":"9d946a60e806e217fa5f56b303307f58b580b8aff6f2cc08c8f6de3ffd1582a3"} Mar 11 13:36:04 crc kubenswrapper[4923]: I0311 13:36:04.262021 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6vcq\" (UniqueName: \"kubernetes.io/projected/6755f934-96d2-4447-b003-aea03c8df883-kube-api-access-r6vcq\") pod \"mariadb-operator-controller-manager-6d9597b6cb-f8lmr\" (UID: \"6755f934-96d2-4447-b003-aea03c8df883\") " pod="openstack-operators/mariadb-operator-controller-manager-6d9597b6cb-f8lmr" Mar 11 13:36:04 crc kubenswrapper[4923]: I0311 13:36:04.262101 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6755f934-96d2-4447-b003-aea03c8df883-apiservice-cert\") pod \"mariadb-operator-controller-manager-6d9597b6cb-f8lmr\" (UID: \"6755f934-96d2-4447-b003-aea03c8df883\") " pod="openstack-operators/mariadb-operator-controller-manager-6d9597b6cb-f8lmr" Mar 11 13:36:04 crc kubenswrapper[4923]: I0311 13:36:04.262211 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6755f934-96d2-4447-b003-aea03c8df883-webhook-cert\") pod \"mariadb-operator-controller-manager-6d9597b6cb-f8lmr\" (UID: \"6755f934-96d2-4447-b003-aea03c8df883\") " pod="openstack-operators/mariadb-operator-controller-manager-6d9597b6cb-f8lmr" Mar 11 13:36:04 crc kubenswrapper[4923]: I0311 13:36:04.362993 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6755f934-96d2-4447-b003-aea03c8df883-webhook-cert\") pod \"mariadb-operator-controller-manager-6d9597b6cb-f8lmr\" (UID: \"6755f934-96d2-4447-b003-aea03c8df883\") " pod="openstack-operators/mariadb-operator-controller-manager-6d9597b6cb-f8lmr" Mar 11 13:36:04 crc kubenswrapper[4923]: I0311 13:36:04.363066 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6vcq\" (UniqueName: \"kubernetes.io/projected/6755f934-96d2-4447-b003-aea03c8df883-kube-api-access-r6vcq\") pod \"mariadb-operator-controller-manager-6d9597b6cb-f8lmr\" (UID: \"6755f934-96d2-4447-b003-aea03c8df883\") " pod="openstack-operators/mariadb-operator-controller-manager-6d9597b6cb-f8lmr" Mar 11 13:36:04 crc kubenswrapper[4923]: I0311 13:36:04.363097 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6755f934-96d2-4447-b003-aea03c8df883-apiservice-cert\") pod \"mariadb-operator-controller-manager-6d9597b6cb-f8lmr\" (UID: \"6755f934-96d2-4447-b003-aea03c8df883\") " pod="openstack-operators/mariadb-operator-controller-manager-6d9597b6cb-f8lmr" Mar 11 13:36:04 crc kubenswrapper[4923]: I0311 13:36:04.367993 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6755f934-96d2-4447-b003-aea03c8df883-webhook-cert\") pod \"mariadb-operator-controller-manager-6d9597b6cb-f8lmr\" (UID: \"6755f934-96d2-4447-b003-aea03c8df883\") " pod="openstack-operators/mariadb-operator-controller-manager-6d9597b6cb-f8lmr" Mar 11 13:36:04 crc kubenswrapper[4923]: I0311 13:36:04.375258 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6755f934-96d2-4447-b003-aea03c8df883-apiservice-cert\") pod \"mariadb-operator-controller-manager-6d9597b6cb-f8lmr\" (UID: \"6755f934-96d2-4447-b003-aea03c8df883\") " pod="openstack-operators/mariadb-operator-controller-manager-6d9597b6cb-f8lmr" Mar 11 13:36:04 crc kubenswrapper[4923]: I0311 13:36:04.378650 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6vcq\" (UniqueName: \"kubernetes.io/projected/6755f934-96d2-4447-b003-aea03c8df883-kube-api-access-r6vcq\") pod \"mariadb-operator-controller-manager-6d9597b6cb-f8lmr\" (UID: \"6755f934-96d2-4447-b003-aea03c8df883\") " pod="openstack-operators/mariadb-operator-controller-manager-6d9597b6cb-f8lmr" Mar 11 13:36:04 crc kubenswrapper[4923]: I0311 13:36:04.457224 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-6d9597b6cb-f8lmr" Mar 11 13:36:04 crc kubenswrapper[4923]: I0311 13:36:04.858439 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-6d9597b6cb-f8lmr"] Mar 11 13:36:05 crc kubenswrapper[4923]: I0311 13:36:05.274932 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6d9597b6cb-f8lmr" event={"ID":"6755f934-96d2-4447-b003-aea03c8df883","Type":"ContainerStarted","Data":"043c72b316e626e4febefa0928cddfcea0477ea8c6bffa353516db019d3bdb38"} Mar 11 13:36:05 crc kubenswrapper[4923]: I0311 13:36:05.555145 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553936-k7s5k" Mar 11 13:36:05 crc kubenswrapper[4923]: I0311 13:36:05.682666 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4dh4d\" (UniqueName: \"kubernetes.io/projected/85eb3ab9-5384-4c51-b20c-66eaa4f44c38-kube-api-access-4dh4d\") pod \"85eb3ab9-5384-4c51-b20c-66eaa4f44c38\" (UID: \"85eb3ab9-5384-4c51-b20c-66eaa4f44c38\") " Mar 11 13:36:05 crc kubenswrapper[4923]: I0311 13:36:05.690595 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85eb3ab9-5384-4c51-b20c-66eaa4f44c38-kube-api-access-4dh4d" (OuterVolumeSpecName: "kube-api-access-4dh4d") pod "85eb3ab9-5384-4c51-b20c-66eaa4f44c38" (UID: "85eb3ab9-5384-4c51-b20c-66eaa4f44c38"). InnerVolumeSpecName "kube-api-access-4dh4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:36:05 crc kubenswrapper[4923]: I0311 13:36:05.785484 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4dh4d\" (UniqueName: \"kubernetes.io/projected/85eb3ab9-5384-4c51-b20c-66eaa4f44c38-kube-api-access-4dh4d\") on node \"crc\" DevicePath \"\"" Mar 11 13:36:06 crc kubenswrapper[4923]: I0311 13:36:06.281048 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553936-k7s5k" event={"ID":"85eb3ab9-5384-4c51-b20c-66eaa4f44c38","Type":"ContainerDied","Data":"117ac95bd0af5823a2239dfecff35274d6288cb3cb25c4934fff31d4158967d8"} Mar 11 13:36:06 crc kubenswrapper[4923]: I0311 13:36:06.281086 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="117ac95bd0af5823a2239dfecff35274d6288cb3cb25c4934fff31d4158967d8" Mar 11 13:36:06 crc kubenswrapper[4923]: I0311 13:36:06.281090 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553936-k7s5k" Mar 11 13:36:06 crc kubenswrapper[4923]: I0311 13:36:06.605487 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29553930-mqfkr"] Mar 11 13:36:06 crc kubenswrapper[4923]: I0311 13:36:06.608914 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29553930-mqfkr"] Mar 11 13:36:07 crc kubenswrapper[4923]: I0311 13:36:07.029003 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fbadffd-f924-474e-af04-11fd967a67a6" path="/var/lib/kubelet/pods/6fbadffd-f924-474e-af04-11fd967a67a6/volumes" Mar 11 13:36:09 crc kubenswrapper[4923]: I0311 13:36:09.304357 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-6d9597b6cb-f8lmr" event={"ID":"6755f934-96d2-4447-b003-aea03c8df883","Type":"ContainerStarted","Data":"4983df6850cc906089d2b358d3c09a6c93a94b0089f31f6269258f3ccb1ba59d"} Mar 11 13:36:09 crc kubenswrapper[4923]: I0311 13:36:09.305586 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-6d9597b6cb-f8lmr" Mar 11 13:36:09 crc kubenswrapper[4923]: I0311 13:36:09.330432 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-6d9597b6cb-f8lmr" podStartSLOduration=2.024435332 podStartE2EDuration="5.330403456s" podCreationTimestamp="2026-03-11 13:36:04 +0000 UTC" firstStartedPulling="2026-03-11 13:36:04.87624274 +0000 UTC m=+812.558144494" lastFinishedPulling="2026-03-11 13:36:08.182210904 +0000 UTC m=+815.864112618" observedRunningTime="2026-03-11 13:36:09.326300651 +0000 UTC m=+817.008202365" watchObservedRunningTime="2026-03-11 13:36:09.330403456 +0000 UTC m=+817.012305210" Mar 11 13:36:14 crc kubenswrapper[4923]: I0311 13:36:14.463289 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-6d9597b6cb-f8lmr" Mar 11 13:36:21 crc kubenswrapper[4923]: I0311 13:36:21.426035 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-p8zpv"] Mar 11 13:36:21 crc kubenswrapper[4923]: E0311 13:36:21.426787 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85eb3ab9-5384-4c51-b20c-66eaa4f44c38" containerName="oc" Mar 11 13:36:21 crc kubenswrapper[4923]: I0311 13:36:21.426803 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="85eb3ab9-5384-4c51-b20c-66eaa4f44c38" containerName="oc" Mar 11 13:36:21 crc kubenswrapper[4923]: I0311 13:36:21.426929 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="85eb3ab9-5384-4c51-b20c-66eaa4f44c38" containerName="oc" Mar 11 13:36:21 crc kubenswrapper[4923]: I0311 13:36:21.427417 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-p8zpv" Mar 11 13:36:21 crc kubenswrapper[4923]: I0311 13:36:21.430704 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-lt5cf" Mar 11 13:36:21 crc kubenswrapper[4923]: I0311 13:36:21.448444 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-p8zpv"] Mar 11 13:36:21 crc kubenswrapper[4923]: I0311 13:36:21.499733 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mk972\" (UniqueName: \"kubernetes.io/projected/656e3de8-2700-4579-ad0f-961699ace32e-kube-api-access-mk972\") pod \"infra-operator-index-p8zpv\" (UID: \"656e3de8-2700-4579-ad0f-961699ace32e\") " pod="openstack-operators/infra-operator-index-p8zpv" Mar 11 13:36:21 crc kubenswrapper[4923]: I0311 13:36:21.600839 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mk972\" (UniqueName: \"kubernetes.io/projected/656e3de8-2700-4579-ad0f-961699ace32e-kube-api-access-mk972\") pod \"infra-operator-index-p8zpv\" (UID: \"656e3de8-2700-4579-ad0f-961699ace32e\") " pod="openstack-operators/infra-operator-index-p8zpv" Mar 11 13:36:21 crc kubenswrapper[4923]: I0311 13:36:21.622167 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mk972\" (UniqueName: \"kubernetes.io/projected/656e3de8-2700-4579-ad0f-961699ace32e-kube-api-access-mk972\") pod \"infra-operator-index-p8zpv\" (UID: \"656e3de8-2700-4579-ad0f-961699ace32e\") " pod="openstack-operators/infra-operator-index-p8zpv" Mar 11 13:36:21 crc kubenswrapper[4923]: I0311 13:36:21.754031 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-p8zpv" Mar 11 13:36:22 crc kubenswrapper[4923]: I0311 13:36:22.000609 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-p8zpv"] Mar 11 13:36:22 crc kubenswrapper[4923]: I0311 13:36:22.403876 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-p8zpv" event={"ID":"656e3de8-2700-4579-ad0f-961699ace32e","Type":"ContainerStarted","Data":"5faf341f04122acaae9dbe07ff4acdee907a320727acb9f1cefc4102f1f85071"} Mar 11 13:36:23 crc kubenswrapper[4923]: I0311 13:36:23.415789 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-p8zpv" event={"ID":"656e3de8-2700-4579-ad0f-961699ace32e","Type":"ContainerStarted","Data":"3990eca9407b96f536e49f21283954a108e2ef9d2026641aa807f78c46385ffe"} Mar 11 13:36:23 crc kubenswrapper[4923]: I0311 13:36:23.440870 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-p8zpv" podStartSLOduration=1.341541506 podStartE2EDuration="2.440851513s" podCreationTimestamp="2026-03-11 13:36:21 +0000 UTC" firstStartedPulling="2026-03-11 13:36:22.010627829 +0000 UTC m=+829.692529543" lastFinishedPulling="2026-03-11 13:36:23.109937826 +0000 UTC m=+830.791839550" observedRunningTime="2026-03-11 13:36:23.437699995 +0000 UTC m=+831.119601749" watchObservedRunningTime="2026-03-11 13:36:23.440851513 +0000 UTC m=+831.122753237" Mar 11 13:36:31 crc kubenswrapper[4923]: I0311 13:36:31.754945 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-p8zpv" Mar 11 13:36:31 crc kubenswrapper[4923]: I0311 13:36:31.755240 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-p8zpv" Mar 11 13:36:31 crc kubenswrapper[4923]: I0311 13:36:31.813649 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-p8zpv" Mar 11 13:36:32 crc kubenswrapper[4923]: I0311 13:36:32.536213 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-p8zpv" Mar 11 13:36:33 crc kubenswrapper[4923]: I0311 13:36:33.679188 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l"] Mar 11 13:36:33 crc kubenswrapper[4923]: I0311 13:36:33.681062 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l" Mar 11 13:36:33 crc kubenswrapper[4923]: I0311 13:36:33.681618 4923 scope.go:117] "RemoveContainer" containerID="a9274dcde3350104b71357d8a252a2cd660c94e7fa939f94b0898d87084ff6f7" Mar 11 13:36:33 crc kubenswrapper[4923]: I0311 13:36:33.685379 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-qkqtx" Mar 11 13:36:33 crc kubenswrapper[4923]: I0311 13:36:33.696208 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l"] Mar 11 13:36:33 crc kubenswrapper[4923]: I0311 13:36:33.770497 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c039997d-8c74-456b-8d4e-250491cff9fc-util\") pod \"d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l\" (UID: \"c039997d-8c74-456b-8d4e-250491cff9fc\") " pod="openstack-operators/d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l" Mar 11 13:36:33 crc kubenswrapper[4923]: I0311 13:36:33.770561 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbc6p\" (UniqueName: \"kubernetes.io/projected/c039997d-8c74-456b-8d4e-250491cff9fc-kube-api-access-kbc6p\") pod \"d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l\" (UID: \"c039997d-8c74-456b-8d4e-250491cff9fc\") " pod="openstack-operators/d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l" Mar 11 13:36:33 crc kubenswrapper[4923]: I0311 13:36:33.770677 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c039997d-8c74-456b-8d4e-250491cff9fc-bundle\") pod \"d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l\" (UID: \"c039997d-8c74-456b-8d4e-250491cff9fc\") " pod="openstack-operators/d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l" Mar 11 13:36:33 crc kubenswrapper[4923]: I0311 13:36:33.871542 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c039997d-8c74-456b-8d4e-250491cff9fc-util\") pod \"d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l\" (UID: \"c039997d-8c74-456b-8d4e-250491cff9fc\") " pod="openstack-operators/d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l" Mar 11 13:36:33 crc kubenswrapper[4923]: I0311 13:36:33.872017 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbc6p\" (UniqueName: \"kubernetes.io/projected/c039997d-8c74-456b-8d4e-250491cff9fc-kube-api-access-kbc6p\") pod \"d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l\" (UID: \"c039997d-8c74-456b-8d4e-250491cff9fc\") " pod="openstack-operators/d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l" Mar 11 13:36:33 crc kubenswrapper[4923]: I0311 13:36:33.872052 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c039997d-8c74-456b-8d4e-250491cff9fc-bundle\") pod \"d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l\" (UID: \"c039997d-8c74-456b-8d4e-250491cff9fc\") " pod="openstack-operators/d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l" Mar 11 13:36:33 crc kubenswrapper[4923]: I0311 13:36:33.872599 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c039997d-8c74-456b-8d4e-250491cff9fc-bundle\") pod \"d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l\" (UID: \"c039997d-8c74-456b-8d4e-250491cff9fc\") " pod="openstack-operators/d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l" Mar 11 13:36:33 crc kubenswrapper[4923]: I0311 13:36:33.873022 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c039997d-8c74-456b-8d4e-250491cff9fc-util\") pod \"d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l\" (UID: \"c039997d-8c74-456b-8d4e-250491cff9fc\") " pod="openstack-operators/d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l" Mar 11 13:36:33 crc kubenswrapper[4923]: I0311 13:36:33.901485 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbc6p\" (UniqueName: \"kubernetes.io/projected/c039997d-8c74-456b-8d4e-250491cff9fc-kube-api-access-kbc6p\") pod \"d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l\" (UID: \"c039997d-8c74-456b-8d4e-250491cff9fc\") " pod="openstack-operators/d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l" Mar 11 13:36:34 crc kubenswrapper[4923]: I0311 13:36:34.039202 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l" Mar 11 13:36:34 crc kubenswrapper[4923]: I0311 13:36:34.555046 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l"] Mar 11 13:36:35 crc kubenswrapper[4923]: I0311 13:36:35.533396 4923 generic.go:334] "Generic (PLEG): container finished" podID="c039997d-8c74-456b-8d4e-250491cff9fc" containerID="6b6009edff26b2810559297833e9fc3ffeb68299adbe3b6579fea70bb5582ab0" exitCode=0 Mar 11 13:36:35 crc kubenswrapper[4923]: I0311 13:36:35.535282 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l" event={"ID":"c039997d-8c74-456b-8d4e-250491cff9fc","Type":"ContainerDied","Data":"6b6009edff26b2810559297833e9fc3ffeb68299adbe3b6579fea70bb5582ab0"} Mar 11 13:36:35 crc kubenswrapper[4923]: I0311 13:36:35.535320 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l" event={"ID":"c039997d-8c74-456b-8d4e-250491cff9fc","Type":"ContainerStarted","Data":"097731830fa4243f4796268842f0e5360b768323dacea27ba62acf5a2798d846"} Mar 11 13:36:37 crc kubenswrapper[4923]: I0311 13:36:37.563462 4923 generic.go:334] "Generic (PLEG): container finished" podID="c039997d-8c74-456b-8d4e-250491cff9fc" containerID="e488200b92176c2e8d642fc9bb96c4b8c51cd6e7fe9c1983fb46daf767ccffb5" exitCode=0 Mar 11 13:36:37 crc kubenswrapper[4923]: I0311 13:36:37.563543 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l" event={"ID":"c039997d-8c74-456b-8d4e-250491cff9fc","Type":"ContainerDied","Data":"e488200b92176c2e8d642fc9bb96c4b8c51cd6e7fe9c1983fb46daf767ccffb5"} Mar 11 13:36:38 crc kubenswrapper[4923]: I0311 13:36:38.577666 4923 generic.go:334] "Generic (PLEG): container finished" podID="c039997d-8c74-456b-8d4e-250491cff9fc" containerID="a77204526a0c3946f6618d4fb06bb0c61c4bca272c976ff727c257a3d67b69f2" exitCode=0 Mar 11 13:36:38 crc kubenswrapper[4923]: I0311 13:36:38.577754 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l" event={"ID":"c039997d-8c74-456b-8d4e-250491cff9fc","Type":"ContainerDied","Data":"a77204526a0c3946f6618d4fb06bb0c61c4bca272c976ff727c257a3d67b69f2"} Mar 11 13:36:39 crc kubenswrapper[4923]: I0311 13:36:39.867102 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l" Mar 11 13:36:39 crc kubenswrapper[4923]: I0311 13:36:39.971651 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbc6p\" (UniqueName: \"kubernetes.io/projected/c039997d-8c74-456b-8d4e-250491cff9fc-kube-api-access-kbc6p\") pod \"c039997d-8c74-456b-8d4e-250491cff9fc\" (UID: \"c039997d-8c74-456b-8d4e-250491cff9fc\") " Mar 11 13:36:39 crc kubenswrapper[4923]: I0311 13:36:39.971764 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c039997d-8c74-456b-8d4e-250491cff9fc-bundle\") pod \"c039997d-8c74-456b-8d4e-250491cff9fc\" (UID: \"c039997d-8c74-456b-8d4e-250491cff9fc\") " Mar 11 13:36:39 crc kubenswrapper[4923]: I0311 13:36:39.971792 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c039997d-8c74-456b-8d4e-250491cff9fc-util\") pod \"c039997d-8c74-456b-8d4e-250491cff9fc\" (UID: \"c039997d-8c74-456b-8d4e-250491cff9fc\") " Mar 11 13:36:39 crc kubenswrapper[4923]: I0311 13:36:39.973391 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c039997d-8c74-456b-8d4e-250491cff9fc-bundle" (OuterVolumeSpecName: "bundle") pod "c039997d-8c74-456b-8d4e-250491cff9fc" (UID: "c039997d-8c74-456b-8d4e-250491cff9fc"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:36:39 crc kubenswrapper[4923]: I0311 13:36:39.984806 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c039997d-8c74-456b-8d4e-250491cff9fc-kube-api-access-kbc6p" (OuterVolumeSpecName: "kube-api-access-kbc6p") pod "c039997d-8c74-456b-8d4e-250491cff9fc" (UID: "c039997d-8c74-456b-8d4e-250491cff9fc"). InnerVolumeSpecName "kube-api-access-kbc6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:36:40 crc kubenswrapper[4923]: I0311 13:36:40.003824 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c039997d-8c74-456b-8d4e-250491cff9fc-util" (OuterVolumeSpecName: "util") pod "c039997d-8c74-456b-8d4e-250491cff9fc" (UID: "c039997d-8c74-456b-8d4e-250491cff9fc"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:36:40 crc kubenswrapper[4923]: I0311 13:36:40.073718 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbc6p\" (UniqueName: \"kubernetes.io/projected/c039997d-8c74-456b-8d4e-250491cff9fc-kube-api-access-kbc6p\") on node \"crc\" DevicePath \"\"" Mar 11 13:36:40 crc kubenswrapper[4923]: I0311 13:36:40.074062 4923 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c039997d-8c74-456b-8d4e-250491cff9fc-bundle\") on node \"crc\" DevicePath \"\"" Mar 11 13:36:40 crc kubenswrapper[4923]: I0311 13:36:40.074289 4923 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c039997d-8c74-456b-8d4e-250491cff9fc-util\") on node \"crc\" DevicePath \"\"" Mar 11 13:36:40 crc kubenswrapper[4923]: I0311 13:36:40.596422 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l" event={"ID":"c039997d-8c74-456b-8d4e-250491cff9fc","Type":"ContainerDied","Data":"097731830fa4243f4796268842f0e5360b768323dacea27ba62acf5a2798d846"} Mar 11 13:36:40 crc kubenswrapper[4923]: I0311 13:36:40.596849 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="097731830fa4243f4796268842f0e5360b768323dacea27ba62acf5a2798d846" Mar 11 13:36:40 crc kubenswrapper[4923]: I0311 13:36:40.596503 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.058107 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Mar 11 13:36:50 crc kubenswrapper[4923]: E0311 13:36:50.058631 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c039997d-8c74-456b-8d4e-250491cff9fc" containerName="extract" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.058647 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="c039997d-8c74-456b-8d4e-250491cff9fc" containerName="extract" Mar 11 13:36:50 crc kubenswrapper[4923]: E0311 13:36:50.058674 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c039997d-8c74-456b-8d4e-250491cff9fc" containerName="util" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.058682 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="c039997d-8c74-456b-8d4e-250491cff9fc" containerName="util" Mar 11 13:36:50 crc kubenswrapper[4923]: E0311 13:36:50.058693 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c039997d-8c74-456b-8d4e-250491cff9fc" containerName="pull" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.058702 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="c039997d-8c74-456b-8d4e-250491cff9fc" containerName="pull" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.058833 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="c039997d-8c74-456b-8d4e-250491cff9fc" containerName="extract" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.059535 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.062715 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"openshift-service-ca.crt" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.063298 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"openstack-scripts" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.063552 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"kube-root-ca.crt" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.063596 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"openstack-config-data" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.063560 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"galera-openstack-dockercfg-wbbqr" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.071674 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.072844 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.087518 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.088610 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.095043 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.104389 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.110257 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.248238 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-2\" (UID: \"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06\") " pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.248421 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/57cc1bb8-4fe5-42af-b6b8-672d0baf4b06-config-data-generated\") pod \"openstack-galera-2\" (UID: \"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06\") " pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.248459 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5b12f995-e853-4d67-a013-beb96a7ac30b-config-data-default\") pod \"openstack-galera-1\" (UID: \"5b12f995-e853-4d67-a013-beb96a7ac30b\") " pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.248486 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/229f08c8-d2cf-46a7-b933-cb5223b654ff-config-data-default\") pod \"openstack-galera-0\" (UID: \"229f08c8-d2cf-46a7-b933-cb5223b654ff\") " pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.248511 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f9q8\" (UniqueName: \"kubernetes.io/projected/57cc1bb8-4fe5-42af-b6b8-672d0baf4b06-kube-api-access-8f9q8\") pod \"openstack-galera-2\" (UID: \"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06\") " pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.248533 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"229f08c8-d2cf-46a7-b933-cb5223b654ff\") " pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.248552 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b12f995-e853-4d67-a013-beb96a7ac30b-operator-scripts\") pod \"openstack-galera-1\" (UID: \"5b12f995-e853-4d67-a013-beb96a7ac30b\") " pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.248566 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57cc1bb8-4fe5-42af-b6b8-672d0baf4b06-operator-scripts\") pod \"openstack-galera-2\" (UID: \"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06\") " pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.248585 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/229f08c8-d2cf-46a7-b933-cb5223b654ff-config-data-generated\") pod \"openstack-galera-0\" (UID: \"229f08c8-d2cf-46a7-b933-cb5223b654ff\") " pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.248601 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ghzv\" (UniqueName: \"kubernetes.io/projected/229f08c8-d2cf-46a7-b933-cb5223b654ff-kube-api-access-5ghzv\") pod \"openstack-galera-0\" (UID: \"229f08c8-d2cf-46a7-b933-cb5223b654ff\") " pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.248620 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmdwt\" (UniqueName: \"kubernetes.io/projected/5b12f995-e853-4d67-a013-beb96a7ac30b-kube-api-access-fmdwt\") pod \"openstack-galera-1\" (UID: \"5b12f995-e853-4d67-a013-beb96a7ac30b\") " pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.248642 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-1\" (UID: \"5b12f995-e853-4d67-a013-beb96a7ac30b\") " pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.248657 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/57cc1bb8-4fe5-42af-b6b8-672d0baf4b06-config-data-default\") pod \"openstack-galera-2\" (UID: \"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06\") " pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.248673 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5b12f995-e853-4d67-a013-beb96a7ac30b-config-data-generated\") pod \"openstack-galera-1\" (UID: \"5b12f995-e853-4d67-a013-beb96a7ac30b\") " pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.248729 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/229f08c8-d2cf-46a7-b933-cb5223b654ff-kolla-config\") pod \"openstack-galera-0\" (UID: \"229f08c8-d2cf-46a7-b933-cb5223b654ff\") " pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.248745 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5b12f995-e853-4d67-a013-beb96a7ac30b-kolla-config\") pod \"openstack-galera-1\" (UID: \"5b12f995-e853-4d67-a013-beb96a7ac30b\") " pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.248811 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/57cc1bb8-4fe5-42af-b6b8-672d0baf4b06-kolla-config\") pod \"openstack-galera-2\" (UID: \"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06\") " pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.248827 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/229f08c8-d2cf-46a7-b933-cb5223b654ff-operator-scripts\") pod \"openstack-galera-0\" (UID: \"229f08c8-d2cf-46a7-b933-cb5223b654ff\") " pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.350352 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f9q8\" (UniqueName: \"kubernetes.io/projected/57cc1bb8-4fe5-42af-b6b8-672d0baf4b06-kube-api-access-8f9q8\") pod \"openstack-galera-2\" (UID: \"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06\") " pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.350624 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"229f08c8-d2cf-46a7-b933-cb5223b654ff\") " pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.350650 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b12f995-e853-4d67-a013-beb96a7ac30b-operator-scripts\") pod \"openstack-galera-1\" (UID: \"5b12f995-e853-4d67-a013-beb96a7ac30b\") " pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.350668 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57cc1bb8-4fe5-42af-b6b8-672d0baf4b06-operator-scripts\") pod \"openstack-galera-2\" (UID: \"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06\") " pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.350688 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/229f08c8-d2cf-46a7-b933-cb5223b654ff-config-data-generated\") pod \"openstack-galera-0\" (UID: \"229f08c8-d2cf-46a7-b933-cb5223b654ff\") " pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.350706 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ghzv\" (UniqueName: \"kubernetes.io/projected/229f08c8-d2cf-46a7-b933-cb5223b654ff-kube-api-access-5ghzv\") pod \"openstack-galera-0\" (UID: \"229f08c8-d2cf-46a7-b933-cb5223b654ff\") " pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.350725 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmdwt\" (UniqueName: \"kubernetes.io/projected/5b12f995-e853-4d67-a013-beb96a7ac30b-kube-api-access-fmdwt\") pod \"openstack-galera-1\" (UID: \"5b12f995-e853-4d67-a013-beb96a7ac30b\") " pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.350748 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-1\" (UID: \"5b12f995-e853-4d67-a013-beb96a7ac30b\") " pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.350765 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/57cc1bb8-4fe5-42af-b6b8-672d0baf4b06-config-data-default\") pod \"openstack-galera-2\" (UID: \"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06\") " pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.350781 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5b12f995-e853-4d67-a013-beb96a7ac30b-config-data-generated\") pod \"openstack-galera-1\" (UID: \"5b12f995-e853-4d67-a013-beb96a7ac30b\") " pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.350804 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/229f08c8-d2cf-46a7-b933-cb5223b654ff-kolla-config\") pod \"openstack-galera-0\" (UID: \"229f08c8-d2cf-46a7-b933-cb5223b654ff\") " pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.350817 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5b12f995-e853-4d67-a013-beb96a7ac30b-kolla-config\") pod \"openstack-galera-1\" (UID: \"5b12f995-e853-4d67-a013-beb96a7ac30b\") " pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.350840 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/57cc1bb8-4fe5-42af-b6b8-672d0baf4b06-kolla-config\") pod \"openstack-galera-2\" (UID: \"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06\") " pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.350855 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/229f08c8-d2cf-46a7-b933-cb5223b654ff-operator-scripts\") pod \"openstack-galera-0\" (UID: \"229f08c8-d2cf-46a7-b933-cb5223b654ff\") " pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.350872 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-2\" (UID: \"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06\") " pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.350889 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/57cc1bb8-4fe5-42af-b6b8-672d0baf4b06-config-data-generated\") pod \"openstack-galera-2\" (UID: \"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06\") " pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.350908 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5b12f995-e853-4d67-a013-beb96a7ac30b-config-data-default\") pod \"openstack-galera-1\" (UID: \"5b12f995-e853-4d67-a013-beb96a7ac30b\") " pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.350924 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/229f08c8-d2cf-46a7-b933-cb5223b654ff-config-data-default\") pod \"openstack-galera-0\" (UID: \"229f08c8-d2cf-46a7-b933-cb5223b654ff\") " pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.351655 4923 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"229f08c8-d2cf-46a7-b933-cb5223b654ff\") device mount path \"/mnt/openstack/pv04\"" pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.351728 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/229f08c8-d2cf-46a7-b933-cb5223b654ff-config-data-default\") pod \"openstack-galera-0\" (UID: \"229f08c8-d2cf-46a7-b933-cb5223b654ff\") " pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.352578 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/57cc1bb8-4fe5-42af-b6b8-672d0baf4b06-config-data-default\") pod \"openstack-galera-2\" (UID: \"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06\") " pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.352736 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/229f08c8-d2cf-46a7-b933-cb5223b654ff-kolla-config\") pod \"openstack-galera-0\" (UID: \"229f08c8-d2cf-46a7-b933-cb5223b654ff\") " pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.352900 4923 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-2\" (UID: \"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06\") device mount path \"/mnt/openstack/pv03\"" pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.353181 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/57cc1bb8-4fe5-42af-b6b8-672d0baf4b06-operator-scripts\") pod \"openstack-galera-2\" (UID: \"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06\") " pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.353458 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/229f08c8-d2cf-46a7-b933-cb5223b654ff-config-data-generated\") pod \"openstack-galera-0\" (UID: \"229f08c8-d2cf-46a7-b933-cb5223b654ff\") " pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.353824 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5b12f995-e853-4d67-a013-beb96a7ac30b-operator-scripts\") pod \"openstack-galera-1\" (UID: \"5b12f995-e853-4d67-a013-beb96a7ac30b\") " pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.353993 4923 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-1\" (UID: \"5b12f995-e853-4d67-a013-beb96a7ac30b\") device mount path \"/mnt/openstack/pv01\"" pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.354062 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5b12f995-e853-4d67-a013-beb96a7ac30b-kolla-config\") pod \"openstack-galera-1\" (UID: \"5b12f995-e853-4d67-a013-beb96a7ac30b\") " pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.354282 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/57cc1bb8-4fe5-42af-b6b8-672d0baf4b06-config-data-generated\") pod \"openstack-galera-2\" (UID: \"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06\") " pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.354388 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5b12f995-e853-4d67-a013-beb96a7ac30b-config-data-generated\") pod \"openstack-galera-1\" (UID: \"5b12f995-e853-4d67-a013-beb96a7ac30b\") " pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.354804 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5b12f995-e853-4d67-a013-beb96a7ac30b-config-data-default\") pod \"openstack-galera-1\" (UID: \"5b12f995-e853-4d67-a013-beb96a7ac30b\") " pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.354969 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/57cc1bb8-4fe5-42af-b6b8-672d0baf4b06-kolla-config\") pod \"openstack-galera-2\" (UID: \"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06\") " pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.355611 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/229f08c8-d2cf-46a7-b933-cb5223b654ff-operator-scripts\") pod \"openstack-galera-0\" (UID: \"229f08c8-d2cf-46a7-b933-cb5223b654ff\") " pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.373015 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"229f08c8-d2cf-46a7-b933-cb5223b654ff\") " pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.374689 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmdwt\" (UniqueName: \"kubernetes.io/projected/5b12f995-e853-4d67-a013-beb96a7ac30b-kube-api-access-fmdwt\") pod \"openstack-galera-1\" (UID: \"5b12f995-e853-4d67-a013-beb96a7ac30b\") " pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.380011 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-1\" (UID: \"5b12f995-e853-4d67-a013-beb96a7ac30b\") " pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.385645 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f9q8\" (UniqueName: \"kubernetes.io/projected/57cc1bb8-4fe5-42af-b6b8-672d0baf4b06-kube-api-access-8f9q8\") pod \"openstack-galera-2\" (UID: \"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06\") " pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.395529 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-galera-2\" (UID: \"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06\") " pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.396234 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ghzv\" (UniqueName: \"kubernetes.io/projected/229f08c8-d2cf-46a7-b933-cb5223b654ff-kube-api-access-5ghzv\") pod \"openstack-galera-0\" (UID: \"229f08c8-d2cf-46a7-b933-cb5223b654ff\") " pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.433732 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.443134 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.452133 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.815853 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-1"] Mar 11 13:36:50 crc kubenswrapper[4923]: W0311 13:36:50.823356 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b12f995_e853_4d67_a013_beb96a7ac30b.slice/crio-7d17ba8efc62edfb900d468fc2231e7b13a323bacb740a2bc572aed7c9f7fdec WatchSource:0}: Error finding container 7d17ba8efc62edfb900d468fc2231e7b13a323bacb740a2bc572aed7c9f7fdec: Status 404 returned error can't find the container with id 7d17ba8efc62edfb900d468fc2231e7b13a323bacb740a2bc572aed7c9f7fdec Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.964858 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-0"] Mar 11 13:36:50 crc kubenswrapper[4923]: W0311 13:36:50.965526 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod229f08c8_d2cf_46a7_b933_cb5223b654ff.slice/crio-f00704eaa75dbed332002bcfb7afbc715a262ce02e83b4dc039507b5c080e122 WatchSource:0}: Error finding container f00704eaa75dbed332002bcfb7afbc715a262ce02e83b4dc039507b5c080e122: Status 404 returned error can't find the container with id f00704eaa75dbed332002bcfb7afbc715a262ce02e83b4dc039507b5c080e122 Mar 11 13:36:50 crc kubenswrapper[4923]: W0311 13:36:50.967273 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod57cc1bb8_4fe5_42af_b6b8_672d0baf4b06.slice/crio-74eb0546c5975f7c68fd3869eb61e1ae3c9c836295e1e35b98758879bfc8b30f WatchSource:0}: Error finding container 74eb0546c5975f7c68fd3869eb61e1ae3c9c836295e1e35b98758879bfc8b30f: Status 404 returned error can't find the container with id 74eb0546c5975f7c68fd3869eb61e1ae3c9c836295e1e35b98758879bfc8b30f Mar 11 13:36:50 crc kubenswrapper[4923]: I0311 13:36:50.970673 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/openstack-galera-2"] Mar 11 13:36:51 crc kubenswrapper[4923]: I0311 13:36:51.617193 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-5b8cc7847-7bm9z"] Mar 11 13:36:51 crc kubenswrapper[4923]: I0311 13:36:51.618633 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5b8cc7847-7bm9z" Mar 11 13:36:51 crc kubenswrapper[4923]: I0311 13:36:51.620935 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Mar 11 13:36:51 crc kubenswrapper[4923]: I0311 13:36:51.622272 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-stxb5" Mar 11 13:36:51 crc kubenswrapper[4923]: I0311 13:36:51.665567 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5b8cc7847-7bm9z"] Mar 11 13:36:51 crc kubenswrapper[4923]: I0311 13:36:51.681584 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"5b12f995-e853-4d67-a013-beb96a7ac30b","Type":"ContainerStarted","Data":"7d17ba8efc62edfb900d468fc2231e7b13a323bacb740a2bc572aed7c9f7fdec"} Mar 11 13:36:51 crc kubenswrapper[4923]: I0311 13:36:51.716883 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06","Type":"ContainerStarted","Data":"74eb0546c5975f7c68fd3869eb61e1ae3c9c836295e1e35b98758879bfc8b30f"} Mar 11 13:36:51 crc kubenswrapper[4923]: I0311 13:36:51.719023 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"229f08c8-d2cf-46a7-b933-cb5223b654ff","Type":"ContainerStarted","Data":"f00704eaa75dbed332002bcfb7afbc715a262ce02e83b4dc039507b5c080e122"} Mar 11 13:36:51 crc kubenswrapper[4923]: I0311 13:36:51.776429 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/88950f18-28e6-4e1f-92a3-c00047b542da-apiservice-cert\") pod \"infra-operator-controller-manager-5b8cc7847-7bm9z\" (UID: \"88950f18-28e6-4e1f-92a3-c00047b542da\") " pod="openstack-operators/infra-operator-controller-manager-5b8cc7847-7bm9z" Mar 11 13:36:51 crc kubenswrapper[4923]: I0311 13:36:51.776778 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqqmk\" (UniqueName: \"kubernetes.io/projected/88950f18-28e6-4e1f-92a3-c00047b542da-kube-api-access-fqqmk\") pod \"infra-operator-controller-manager-5b8cc7847-7bm9z\" (UID: \"88950f18-28e6-4e1f-92a3-c00047b542da\") " pod="openstack-operators/infra-operator-controller-manager-5b8cc7847-7bm9z" Mar 11 13:36:51 crc kubenswrapper[4923]: I0311 13:36:51.776851 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/88950f18-28e6-4e1f-92a3-c00047b542da-webhook-cert\") pod \"infra-operator-controller-manager-5b8cc7847-7bm9z\" (UID: \"88950f18-28e6-4e1f-92a3-c00047b542da\") " pod="openstack-operators/infra-operator-controller-manager-5b8cc7847-7bm9z" Mar 11 13:36:51 crc kubenswrapper[4923]: I0311 13:36:51.878241 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqqmk\" (UniqueName: \"kubernetes.io/projected/88950f18-28e6-4e1f-92a3-c00047b542da-kube-api-access-fqqmk\") pod \"infra-operator-controller-manager-5b8cc7847-7bm9z\" (UID: \"88950f18-28e6-4e1f-92a3-c00047b542da\") " pod="openstack-operators/infra-operator-controller-manager-5b8cc7847-7bm9z" Mar 11 13:36:51 crc kubenswrapper[4923]: I0311 13:36:51.878315 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/88950f18-28e6-4e1f-92a3-c00047b542da-webhook-cert\") pod \"infra-operator-controller-manager-5b8cc7847-7bm9z\" (UID: \"88950f18-28e6-4e1f-92a3-c00047b542da\") " pod="openstack-operators/infra-operator-controller-manager-5b8cc7847-7bm9z" Mar 11 13:36:51 crc kubenswrapper[4923]: I0311 13:36:51.878374 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/88950f18-28e6-4e1f-92a3-c00047b542da-apiservice-cert\") pod \"infra-operator-controller-manager-5b8cc7847-7bm9z\" (UID: \"88950f18-28e6-4e1f-92a3-c00047b542da\") " pod="openstack-operators/infra-operator-controller-manager-5b8cc7847-7bm9z" Mar 11 13:36:51 crc kubenswrapper[4923]: I0311 13:36:51.886919 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/88950f18-28e6-4e1f-92a3-c00047b542da-apiservice-cert\") pod \"infra-operator-controller-manager-5b8cc7847-7bm9z\" (UID: \"88950f18-28e6-4e1f-92a3-c00047b542da\") " pod="openstack-operators/infra-operator-controller-manager-5b8cc7847-7bm9z" Mar 11 13:36:51 crc kubenswrapper[4923]: I0311 13:36:51.888925 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/88950f18-28e6-4e1f-92a3-c00047b542da-webhook-cert\") pod \"infra-operator-controller-manager-5b8cc7847-7bm9z\" (UID: \"88950f18-28e6-4e1f-92a3-c00047b542da\") " pod="openstack-operators/infra-operator-controller-manager-5b8cc7847-7bm9z" Mar 11 13:36:51 crc kubenswrapper[4923]: I0311 13:36:51.904001 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqqmk\" (UniqueName: \"kubernetes.io/projected/88950f18-28e6-4e1f-92a3-c00047b542da-kube-api-access-fqqmk\") pod \"infra-operator-controller-manager-5b8cc7847-7bm9z\" (UID: \"88950f18-28e6-4e1f-92a3-c00047b542da\") " pod="openstack-operators/infra-operator-controller-manager-5b8cc7847-7bm9z" Mar 11 13:36:51 crc kubenswrapper[4923]: I0311 13:36:51.953026 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5b8cc7847-7bm9z" Mar 11 13:36:52 crc kubenswrapper[4923]: I0311 13:36:52.246226 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5b8cc7847-7bm9z"] Mar 11 13:36:52 crc kubenswrapper[4923]: W0311 13:36:52.273312 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88950f18_28e6_4e1f_92a3_c00047b542da.slice/crio-5916d58320a64cbc2708fb9e6494eb69793b292063ad038c93b4c642d97122a3 WatchSource:0}: Error finding container 5916d58320a64cbc2708fb9e6494eb69793b292063ad038c93b4c642d97122a3: Status 404 returned error can't find the container with id 5916d58320a64cbc2708fb9e6494eb69793b292063ad038c93b4c642d97122a3 Mar 11 13:36:52 crc kubenswrapper[4923]: I0311 13:36:52.725741 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5b8cc7847-7bm9z" event={"ID":"88950f18-28e6-4e1f-92a3-c00047b542da","Type":"ContainerStarted","Data":"5916d58320a64cbc2708fb9e6494eb69793b292063ad038c93b4c642d97122a3"} Mar 11 13:37:00 crc kubenswrapper[4923]: I0311 13:37:00.777355 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5b8cc7847-7bm9z" event={"ID":"88950f18-28e6-4e1f-92a3-c00047b542da","Type":"ContainerStarted","Data":"d879e861c43a818e1a3e1a12d4c7af5cda0f3710af8eae3be3af56271f345b4a"} Mar 11 13:37:00 crc kubenswrapper[4923]: I0311 13:37:00.777914 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5b8cc7847-7bm9z" Mar 11 13:37:00 crc kubenswrapper[4923]: I0311 13:37:00.779109 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"229f08c8-d2cf-46a7-b933-cb5223b654ff","Type":"ContainerStarted","Data":"91fa6e9fa9c36bef1ca159d8ff6c280614cb2f1bdfc6f8ad369c8f41a01583ca"} Mar 11 13:37:00 crc kubenswrapper[4923]: I0311 13:37:00.780253 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"5b12f995-e853-4d67-a013-beb96a7ac30b","Type":"ContainerStarted","Data":"f24a70c55d29472c501cfa7022034f97bd7ed2054b1e797a5f761f97f7186c43"} Mar 11 13:37:00 crc kubenswrapper[4923]: I0311 13:37:00.781661 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06","Type":"ContainerStarted","Data":"7a70890b1d7074b351425a51131e8b5e3034d49243ff49a67b563e6470021509"} Mar 11 13:37:00 crc kubenswrapper[4923]: I0311 13:37:00.811399 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-5b8cc7847-7bm9z" podStartSLOduration=2.544483148 podStartE2EDuration="9.811373654s" podCreationTimestamp="2026-03-11 13:36:51 +0000 UTC" firstStartedPulling="2026-03-11 13:36:52.286662864 +0000 UTC m=+859.968564578" lastFinishedPulling="2026-03-11 13:36:59.55355336 +0000 UTC m=+867.235455084" observedRunningTime="2026-03-11 13:37:00.808492494 +0000 UTC m=+868.490394218" watchObservedRunningTime="2026-03-11 13:37:00.811373654 +0000 UTC m=+868.493275368" Mar 11 13:37:03 crc kubenswrapper[4923]: I0311 13:37:03.809662 4923 generic.go:334] "Generic (PLEG): container finished" podID="5b12f995-e853-4d67-a013-beb96a7ac30b" containerID="f24a70c55d29472c501cfa7022034f97bd7ed2054b1e797a5f761f97f7186c43" exitCode=0 Mar 11 13:37:03 crc kubenswrapper[4923]: I0311 13:37:03.809949 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"5b12f995-e853-4d67-a013-beb96a7ac30b","Type":"ContainerDied","Data":"f24a70c55d29472c501cfa7022034f97bd7ed2054b1e797a5f761f97f7186c43"} Mar 11 13:37:03 crc kubenswrapper[4923]: I0311 13:37:03.811978 4923 generic.go:334] "Generic (PLEG): container finished" podID="57cc1bb8-4fe5-42af-b6b8-672d0baf4b06" containerID="7a70890b1d7074b351425a51131e8b5e3034d49243ff49a67b563e6470021509" exitCode=0 Mar 11 13:37:03 crc kubenswrapper[4923]: I0311 13:37:03.812002 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06","Type":"ContainerDied","Data":"7a70890b1d7074b351425a51131e8b5e3034d49243ff49a67b563e6470021509"} Mar 11 13:37:03 crc kubenswrapper[4923]: I0311 13:37:03.813744 4923 generic.go:334] "Generic (PLEG): container finished" podID="229f08c8-d2cf-46a7-b933-cb5223b654ff" containerID="91fa6e9fa9c36bef1ca159d8ff6c280614cb2f1bdfc6f8ad369c8f41a01583ca" exitCode=0 Mar 11 13:37:03 crc kubenswrapper[4923]: I0311 13:37:03.813785 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"229f08c8-d2cf-46a7-b933-cb5223b654ff","Type":"ContainerDied","Data":"91fa6e9fa9c36bef1ca159d8ff6c280614cb2f1bdfc6f8ad369c8f41a01583ca"} Mar 11 13:37:04 crc kubenswrapper[4923]: I0311 13:37:04.822928 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-2" event={"ID":"57cc1bb8-4fe5-42af-b6b8-672d0baf4b06","Type":"ContainerStarted","Data":"087fa0fb56515e05028bf2ac8d207255580d4561907514277fa3534575f47466"} Mar 11 13:37:04 crc kubenswrapper[4923]: I0311 13:37:04.826256 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-0" event={"ID":"229f08c8-d2cf-46a7-b933-cb5223b654ff","Type":"ContainerStarted","Data":"fd806f5d14f65c71f5480047f93b009551f2a64fb170e755aef0c04170e81712"} Mar 11 13:37:04 crc kubenswrapper[4923]: I0311 13:37:04.829417 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/openstack-galera-1" event={"ID":"5b12f995-e853-4d67-a013-beb96a7ac30b","Type":"ContainerStarted","Data":"f21df1a79b1488e1fc61e121a0a2add7b591357ebb6aaa69efab8ef27761a858"} Mar 11 13:37:04 crc kubenswrapper[4923]: I0311 13:37:04.984237 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/openstack-galera-2" podStartSLOduration=7.351097655 podStartE2EDuration="15.98421282s" podCreationTimestamp="2026-03-11 13:36:49 +0000 UTC" firstStartedPulling="2026-03-11 13:36:50.969640004 +0000 UTC m=+858.651541728" lastFinishedPulling="2026-03-11 13:36:59.602755179 +0000 UTC m=+867.284656893" observedRunningTime="2026-03-11 13:37:04.979606934 +0000 UTC m=+872.661508658" watchObservedRunningTime="2026-03-11 13:37:04.98421282 +0000 UTC m=+872.666114554" Mar 11 13:37:05 crc kubenswrapper[4923]: I0311 13:37:05.036768 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/openstack-galera-1" podStartSLOduration=7.179339269 podStartE2EDuration="16.036743752s" podCreationTimestamp="2026-03-11 13:36:49 +0000 UTC" firstStartedPulling="2026-03-11 13:36:50.825295665 +0000 UTC m=+858.507197379" lastFinishedPulling="2026-03-11 13:36:59.682700148 +0000 UTC m=+867.364601862" observedRunningTime="2026-03-11 13:37:05.004958734 +0000 UTC m=+872.686860448" watchObservedRunningTime="2026-03-11 13:37:05.036743752 +0000 UTC m=+872.718645476" Mar 11 13:37:05 crc kubenswrapper[4923]: I0311 13:37:05.042949 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/openstack-galera-0" podStartSLOduration=7.4099083199999995 podStartE2EDuration="16.042923233s" podCreationTimestamp="2026-03-11 13:36:49 +0000 UTC" firstStartedPulling="2026-03-11 13:36:50.970619081 +0000 UTC m=+858.652520815" lastFinishedPulling="2026-03-11 13:36:59.603634014 +0000 UTC m=+867.285535728" observedRunningTime="2026-03-11 13:37:05.032418883 +0000 UTC m=+872.714320617" watchObservedRunningTime="2026-03-11 13:37:05.042923233 +0000 UTC m=+872.724824977" Mar 11 13:37:10 crc kubenswrapper[4923]: I0311 13:37:10.434237 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:37:10 crc kubenswrapper[4923]: I0311 13:37:10.435676 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:37:10 crc kubenswrapper[4923]: I0311 13:37:10.446502 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:37:10 crc kubenswrapper[4923]: I0311 13:37:10.446713 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:37:10 crc kubenswrapper[4923]: I0311 13:37:10.452249 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:37:10 crc kubenswrapper[4923]: I0311 13:37:10.452290 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:37:11 crc kubenswrapper[4923]: I0311 13:37:11.932473 4923 patch_prober.go:28] interesting pod/machine-config-daemon-fsfnw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 11 13:37:11 crc kubenswrapper[4923]: I0311 13:37:11.932545 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 11 13:37:11 crc kubenswrapper[4923]: I0311 13:37:11.958054 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5b8cc7847-7bm9z" Mar 11 13:37:14 crc kubenswrapper[4923]: I0311 13:37:14.533950 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:37:14 crc kubenswrapper[4923]: I0311 13:37:14.626531 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/openstack-galera-2" Mar 11 13:37:14 crc kubenswrapper[4923]: E0311 13:37:14.914733 4923 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.111:59334->38.102.83.111:44013: write tcp 38.102.83.111:59334->38.102.83.111:44013: write: broken pipe Mar 11 13:37:15 crc kubenswrapper[4923]: I0311 13:37:15.823933 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/memcached-0"] Mar 11 13:37:15 crc kubenswrapper[4923]: I0311 13:37:15.825222 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/memcached-0" Mar 11 13:37:15 crc kubenswrapper[4923]: I0311 13:37:15.827251 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"memcached-memcached-dockercfg-dlvhj" Mar 11 13:37:15 crc kubenswrapper[4923]: I0311 13:37:15.827561 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"memcached-config-data" Mar 11 13:37:15 crc kubenswrapper[4923]: I0311 13:37:15.838062 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/memcached-0"] Mar 11 13:37:15 crc kubenswrapper[4923]: I0311 13:37:15.895234 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6ccfe8a9-edc8-41a2-af22-d55181b3dba1-config-data\") pod \"memcached-0\" (UID: \"6ccfe8a9-edc8-41a2-af22-d55181b3dba1\") " pod="cinder-kuttl-tests/memcached-0" Mar 11 13:37:15 crc kubenswrapper[4923]: I0311 13:37:15.895290 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vqkk\" (UniqueName: \"kubernetes.io/projected/6ccfe8a9-edc8-41a2-af22-d55181b3dba1-kube-api-access-8vqkk\") pod \"memcached-0\" (UID: \"6ccfe8a9-edc8-41a2-af22-d55181b3dba1\") " pod="cinder-kuttl-tests/memcached-0" Mar 11 13:37:15 crc kubenswrapper[4923]: I0311 13:37:15.895313 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6ccfe8a9-edc8-41a2-af22-d55181b3dba1-kolla-config\") pod \"memcached-0\" (UID: \"6ccfe8a9-edc8-41a2-af22-d55181b3dba1\") " pod="cinder-kuttl-tests/memcached-0" Mar 11 13:37:15 crc kubenswrapper[4923]: I0311 13:37:15.996872 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6ccfe8a9-edc8-41a2-af22-d55181b3dba1-config-data\") pod \"memcached-0\" (UID: \"6ccfe8a9-edc8-41a2-af22-d55181b3dba1\") " pod="cinder-kuttl-tests/memcached-0" Mar 11 13:37:15 crc kubenswrapper[4923]: I0311 13:37:15.996926 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vqkk\" (UniqueName: \"kubernetes.io/projected/6ccfe8a9-edc8-41a2-af22-d55181b3dba1-kube-api-access-8vqkk\") pod \"memcached-0\" (UID: \"6ccfe8a9-edc8-41a2-af22-d55181b3dba1\") " pod="cinder-kuttl-tests/memcached-0" Mar 11 13:37:15 crc kubenswrapper[4923]: I0311 13:37:15.996945 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6ccfe8a9-edc8-41a2-af22-d55181b3dba1-kolla-config\") pod \"memcached-0\" (UID: \"6ccfe8a9-edc8-41a2-af22-d55181b3dba1\") " pod="cinder-kuttl-tests/memcached-0" Mar 11 13:37:15 crc kubenswrapper[4923]: I0311 13:37:15.998021 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6ccfe8a9-edc8-41a2-af22-d55181b3dba1-kolla-config\") pod \"memcached-0\" (UID: \"6ccfe8a9-edc8-41a2-af22-d55181b3dba1\") " pod="cinder-kuttl-tests/memcached-0" Mar 11 13:37:15 crc kubenswrapper[4923]: I0311 13:37:15.998246 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6ccfe8a9-edc8-41a2-af22-d55181b3dba1-config-data\") pod \"memcached-0\" (UID: \"6ccfe8a9-edc8-41a2-af22-d55181b3dba1\") " pod="cinder-kuttl-tests/memcached-0" Mar 11 13:37:16 crc kubenswrapper[4923]: I0311 13:37:16.025424 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vqkk\" (UniqueName: \"kubernetes.io/projected/6ccfe8a9-edc8-41a2-af22-d55181b3dba1-kube-api-access-8vqkk\") pod \"memcached-0\" (UID: \"6ccfe8a9-edc8-41a2-af22-d55181b3dba1\") " pod="cinder-kuttl-tests/memcached-0" Mar 11 13:37:16 crc kubenswrapper[4923]: I0311 13:37:16.207967 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/memcached-0" Mar 11 13:37:16 crc kubenswrapper[4923]: I0311 13:37:16.442703 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/memcached-0"] Mar 11 13:37:16 crc kubenswrapper[4923]: I0311 13:37:16.907142 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/memcached-0" event={"ID":"6ccfe8a9-edc8-41a2-af22-d55181b3dba1","Type":"ContainerStarted","Data":"e5e3a8c296b85497443840b8f7f908fe5ca3d7cc1ae4ca3063e4d86ee37f82e3"} Mar 11 13:37:18 crc kubenswrapper[4923]: I0311 13:37:18.601661 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-64jx7"] Mar 11 13:37:18 crc kubenswrapper[4923]: I0311 13:37:18.603382 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-64jx7" Mar 11 13:37:18 crc kubenswrapper[4923]: I0311 13:37:18.606704 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-n24jj" Mar 11 13:37:18 crc kubenswrapper[4923]: I0311 13:37:18.606869 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-64jx7"] Mar 11 13:37:18 crc kubenswrapper[4923]: I0311 13:37:18.738273 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbg4d\" (UniqueName: \"kubernetes.io/projected/aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c-kube-api-access-cbg4d\") pod \"rabbitmq-cluster-operator-index-64jx7\" (UID: \"aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c\") " pod="openstack-operators/rabbitmq-cluster-operator-index-64jx7" Mar 11 13:37:18 crc kubenswrapper[4923]: I0311 13:37:18.840429 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbg4d\" (UniqueName: \"kubernetes.io/projected/aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c-kube-api-access-cbg4d\") pod \"rabbitmq-cluster-operator-index-64jx7\" (UID: \"aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c\") " pod="openstack-operators/rabbitmq-cluster-operator-index-64jx7" Mar 11 13:37:18 crc kubenswrapper[4923]: I0311 13:37:18.859794 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbg4d\" (UniqueName: \"kubernetes.io/projected/aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c-kube-api-access-cbg4d\") pod \"rabbitmq-cluster-operator-index-64jx7\" (UID: \"aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c\") " pod="openstack-operators/rabbitmq-cluster-operator-index-64jx7" Mar 11 13:37:18 crc kubenswrapper[4923]: I0311 13:37:18.945687 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-64jx7" Mar 11 13:37:19 crc kubenswrapper[4923]: I0311 13:37:19.131612 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/root-account-create-update-hps8z"] Mar 11 13:37:19 crc kubenswrapper[4923]: I0311 13:37:19.133104 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/root-account-create-update-hps8z" Mar 11 13:37:19 crc kubenswrapper[4923]: I0311 13:37:19.151677 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"openstack-mariadb-root-db-secret" Mar 11 13:37:19 crc kubenswrapper[4923]: I0311 13:37:19.159490 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/root-account-create-update-hps8z"] Mar 11 13:37:19 crc kubenswrapper[4923]: I0311 13:37:19.258137 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m7g4\" (UniqueName: \"kubernetes.io/projected/1da28d4d-80b8-4a02-8009-0b6032044448-kube-api-access-6m7g4\") pod \"root-account-create-update-hps8z\" (UID: \"1da28d4d-80b8-4a02-8009-0b6032044448\") " pod="cinder-kuttl-tests/root-account-create-update-hps8z" Mar 11 13:37:19 crc kubenswrapper[4923]: I0311 13:37:19.258204 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1da28d4d-80b8-4a02-8009-0b6032044448-operator-scripts\") pod \"root-account-create-update-hps8z\" (UID: \"1da28d4d-80b8-4a02-8009-0b6032044448\") " pod="cinder-kuttl-tests/root-account-create-update-hps8z" Mar 11 13:37:19 crc kubenswrapper[4923]: I0311 13:37:19.358960 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1da28d4d-80b8-4a02-8009-0b6032044448-operator-scripts\") pod \"root-account-create-update-hps8z\" (UID: \"1da28d4d-80b8-4a02-8009-0b6032044448\") " pod="cinder-kuttl-tests/root-account-create-update-hps8z" Mar 11 13:37:19 crc kubenswrapper[4923]: I0311 13:37:19.359298 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m7g4\" (UniqueName: \"kubernetes.io/projected/1da28d4d-80b8-4a02-8009-0b6032044448-kube-api-access-6m7g4\") pod \"root-account-create-update-hps8z\" (UID: \"1da28d4d-80b8-4a02-8009-0b6032044448\") " pod="cinder-kuttl-tests/root-account-create-update-hps8z" Mar 11 13:37:19 crc kubenswrapper[4923]: I0311 13:37:19.359846 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1da28d4d-80b8-4a02-8009-0b6032044448-operator-scripts\") pod \"root-account-create-update-hps8z\" (UID: \"1da28d4d-80b8-4a02-8009-0b6032044448\") " pod="cinder-kuttl-tests/root-account-create-update-hps8z" Mar 11 13:37:19 crc kubenswrapper[4923]: I0311 13:37:19.374971 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m7g4\" (UniqueName: \"kubernetes.io/projected/1da28d4d-80b8-4a02-8009-0b6032044448-kube-api-access-6m7g4\") pod \"root-account-create-update-hps8z\" (UID: \"1da28d4d-80b8-4a02-8009-0b6032044448\") " pod="cinder-kuttl-tests/root-account-create-update-hps8z" Mar 11 13:37:19 crc kubenswrapper[4923]: I0311 13:37:19.449171 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/root-account-create-update-hps8z" Mar 11 13:37:19 crc kubenswrapper[4923]: I0311 13:37:19.820946 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-64jx7"] Mar 11 13:37:19 crc kubenswrapper[4923]: W0311 13:37:19.829598 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaaad4f2b_bbc4_4f30_807a_c1f1a0ecaa3c.slice/crio-151179b3514d2837b6914072e9eea3fc330257a711379cc1e624133f329beadb WatchSource:0}: Error finding container 151179b3514d2837b6914072e9eea3fc330257a711379cc1e624133f329beadb: Status 404 returned error can't find the container with id 151179b3514d2837b6914072e9eea3fc330257a711379cc1e624133f329beadb Mar 11 13:37:19 crc kubenswrapper[4923]: I0311 13:37:19.920130 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/root-account-create-update-hps8z"] Mar 11 13:37:19 crc kubenswrapper[4923]: W0311 13:37:19.939250 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1da28d4d_80b8_4a02_8009_0b6032044448.slice/crio-6babf62a8ed22ca39dd6add492047ade86104b102df705db0ad86fbbf199288c WatchSource:0}: Error finding container 6babf62a8ed22ca39dd6add492047ade86104b102df705db0ad86fbbf199288c: Status 404 returned error can't find the container with id 6babf62a8ed22ca39dd6add492047ade86104b102df705db0ad86fbbf199288c Mar 11 13:37:19 crc kubenswrapper[4923]: I0311 13:37:19.944928 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-64jx7" event={"ID":"aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c","Type":"ContainerStarted","Data":"151179b3514d2837b6914072e9eea3fc330257a711379cc1e624133f329beadb"} Mar 11 13:37:20 crc kubenswrapper[4923]: I0311 13:37:20.557794 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="cinder-kuttl-tests/openstack-galera-2" podUID="57cc1bb8-4fe5-42af-b6b8-672d0baf4b06" containerName="galera" probeResult="failure" output=< Mar 11 13:37:20 crc kubenswrapper[4923]: wsrep_local_state_comment (Donor/Desynced) differs from Synced Mar 11 13:37:20 crc kubenswrapper[4923]: > Mar 11 13:37:20 crc kubenswrapper[4923]: I0311 13:37:20.954163 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/root-account-create-update-hps8z" event={"ID":"1da28d4d-80b8-4a02-8009-0b6032044448","Type":"ContainerStarted","Data":"c773a542a34bd7b1ca88cfa25860316a4806ff4032287ddf0e84a23098fb6f22"} Mar 11 13:37:20 crc kubenswrapper[4923]: I0311 13:37:20.954213 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/root-account-create-update-hps8z" event={"ID":"1da28d4d-80b8-4a02-8009-0b6032044448","Type":"ContainerStarted","Data":"6babf62a8ed22ca39dd6add492047ade86104b102df705db0ad86fbbf199288c"} Mar 11 13:37:20 crc kubenswrapper[4923]: I0311 13:37:20.972123 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/root-account-create-update-hps8z" podStartSLOduration=1.9721049019999999 podStartE2EDuration="1.972104902s" podCreationTimestamp="2026-03-11 13:37:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:37:20.96845673 +0000 UTC m=+888.650358444" watchObservedRunningTime="2026-03-11 13:37:20.972104902 +0000 UTC m=+888.654006616" Mar 11 13:37:22 crc kubenswrapper[4923]: I0311 13:37:22.968945 4923 generic.go:334] "Generic (PLEG): container finished" podID="1da28d4d-80b8-4a02-8009-0b6032044448" containerID="c773a542a34bd7b1ca88cfa25860316a4806ff4032287ddf0e84a23098fb6f22" exitCode=0 Mar 11 13:37:22 crc kubenswrapper[4923]: I0311 13:37:22.969062 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/root-account-create-update-hps8z" event={"ID":"1da28d4d-80b8-4a02-8009-0b6032044448","Type":"ContainerDied","Data":"c773a542a34bd7b1ca88cfa25860316a4806ff4032287ddf0e84a23098fb6f22"} Mar 11 13:37:23 crc kubenswrapper[4923]: I0311 13:37:23.187292 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-64jx7"] Mar 11 13:37:23 crc kubenswrapper[4923]: I0311 13:37:23.810679 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-sd5l9"] Mar 11 13:37:23 crc kubenswrapper[4923]: I0311 13:37:23.819783 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-sd5l9" Mar 11 13:37:23 crc kubenswrapper[4923]: I0311 13:37:23.829206 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-sd5l9"] Mar 11 13:37:23 crc kubenswrapper[4923]: I0311 13:37:23.925654 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsdcn\" (UniqueName: \"kubernetes.io/projected/9e845d8a-4bf4-457c-a8c3-871ed232846d-kube-api-access-hsdcn\") pod \"rabbitmq-cluster-operator-index-sd5l9\" (UID: \"9e845d8a-4bf4-457c-a8c3-871ed232846d\") " pod="openstack-operators/rabbitmq-cluster-operator-index-sd5l9" Mar 11 13:37:24 crc kubenswrapper[4923]: I0311 13:37:24.027605 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsdcn\" (UniqueName: \"kubernetes.io/projected/9e845d8a-4bf4-457c-a8c3-871ed232846d-kube-api-access-hsdcn\") pod \"rabbitmq-cluster-operator-index-sd5l9\" (UID: \"9e845d8a-4bf4-457c-a8c3-871ed232846d\") " pod="openstack-operators/rabbitmq-cluster-operator-index-sd5l9" Mar 11 13:37:24 crc kubenswrapper[4923]: I0311 13:37:24.049248 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsdcn\" (UniqueName: \"kubernetes.io/projected/9e845d8a-4bf4-457c-a8c3-871ed232846d-kube-api-access-hsdcn\") pod \"rabbitmq-cluster-operator-index-sd5l9\" (UID: \"9e845d8a-4bf4-457c-a8c3-871ed232846d\") " pod="openstack-operators/rabbitmq-cluster-operator-index-sd5l9" Mar 11 13:37:24 crc kubenswrapper[4923]: I0311 13:37:24.147482 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-sd5l9" Mar 11 13:37:25 crc kubenswrapper[4923]: I0311 13:37:25.655158 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:37:25 crc kubenswrapper[4923]: I0311 13:37:25.727200 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/openstack-galera-1" Mar 11 13:37:26 crc kubenswrapper[4923]: I0311 13:37:26.894479 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/root-account-create-update-hps8z" Mar 11 13:37:26 crc kubenswrapper[4923]: I0311 13:37:26.970405 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1da28d4d-80b8-4a02-8009-0b6032044448-operator-scripts\") pod \"1da28d4d-80b8-4a02-8009-0b6032044448\" (UID: \"1da28d4d-80b8-4a02-8009-0b6032044448\") " Mar 11 13:37:26 crc kubenswrapper[4923]: I0311 13:37:26.970547 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6m7g4\" (UniqueName: \"kubernetes.io/projected/1da28d4d-80b8-4a02-8009-0b6032044448-kube-api-access-6m7g4\") pod \"1da28d4d-80b8-4a02-8009-0b6032044448\" (UID: \"1da28d4d-80b8-4a02-8009-0b6032044448\") " Mar 11 13:37:26 crc kubenswrapper[4923]: I0311 13:37:26.970901 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1da28d4d-80b8-4a02-8009-0b6032044448-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1da28d4d-80b8-4a02-8009-0b6032044448" (UID: "1da28d4d-80b8-4a02-8009-0b6032044448"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:37:26 crc kubenswrapper[4923]: I0311 13:37:26.986246 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1da28d4d-80b8-4a02-8009-0b6032044448-kube-api-access-6m7g4" (OuterVolumeSpecName: "kube-api-access-6m7g4") pod "1da28d4d-80b8-4a02-8009-0b6032044448" (UID: "1da28d4d-80b8-4a02-8009-0b6032044448"). InnerVolumeSpecName "kube-api-access-6m7g4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:37:26 crc kubenswrapper[4923]: I0311 13:37:26.993181 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/root-account-create-update-hps8z" event={"ID":"1da28d4d-80b8-4a02-8009-0b6032044448","Type":"ContainerDied","Data":"6babf62a8ed22ca39dd6add492047ade86104b102df705db0ad86fbbf199288c"} Mar 11 13:37:26 crc kubenswrapper[4923]: I0311 13:37:26.993217 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6babf62a8ed22ca39dd6add492047ade86104b102df705db0ad86fbbf199288c" Mar 11 13:37:26 crc kubenswrapper[4923]: I0311 13:37:26.993245 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/root-account-create-update-hps8z" Mar 11 13:37:27 crc kubenswrapper[4923]: I0311 13:37:27.072500 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6m7g4\" (UniqueName: \"kubernetes.io/projected/1da28d4d-80b8-4a02-8009-0b6032044448-kube-api-access-6m7g4\") on node \"crc\" DevicePath \"\"" Mar 11 13:37:27 crc kubenswrapper[4923]: I0311 13:37:27.072544 4923 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1da28d4d-80b8-4a02-8009-0b6032044448-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 11 13:37:28 crc kubenswrapper[4923]: I0311 13:37:28.628528 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-sd5l9"] Mar 11 13:37:28 crc kubenswrapper[4923]: I0311 13:37:28.782598 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:37:28 crc kubenswrapper[4923]: I0311 13:37:28.886949 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/openstack-galera-0" Mar 11 13:37:29 crc kubenswrapper[4923]: I0311 13:37:29.006632 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-sd5l9" event={"ID":"9e845d8a-4bf4-457c-a8c3-871ed232846d","Type":"ContainerStarted","Data":"af78a4f2c9d47088f10ac21e80b0d4598feae2f54802caf0dd3ca3e8222be1f9"} Mar 11 13:37:29 crc kubenswrapper[4923]: I0311 13:37:29.007927 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/memcached-0" event={"ID":"6ccfe8a9-edc8-41a2-af22-d55181b3dba1","Type":"ContainerStarted","Data":"763bb71dea0ebd7bb4a440896f15d5d73c474d204c49fc6bb9359bb44418df12"} Mar 11 13:37:29 crc kubenswrapper[4923]: I0311 13:37:29.008048 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/memcached-0" Mar 11 13:37:29 crc kubenswrapper[4923]: I0311 13:37:29.019209 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/rabbitmq-cluster-operator-index-64jx7" podUID="aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c" containerName="registry-server" containerID="cri-o://43dfddabcd0b748bf7114df586ef20f07bb807379213eb47ba0114b5c88bfc6d" gracePeriod=2 Mar 11 13:37:29 crc kubenswrapper[4923]: I0311 13:37:29.019211 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-64jx7" event={"ID":"aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c","Type":"ContainerStarted","Data":"43dfddabcd0b748bf7114df586ef20f07bb807379213eb47ba0114b5c88bfc6d"} Mar 11 13:37:29 crc kubenswrapper[4923]: I0311 13:37:29.038966 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/memcached-0" podStartSLOduration=2.247598712 podStartE2EDuration="14.03889829s" podCreationTimestamp="2026-03-11 13:37:15 +0000 UTC" firstStartedPulling="2026-03-11 13:37:16.456067012 +0000 UTC m=+884.137968726" lastFinishedPulling="2026-03-11 13:37:28.24736658 +0000 UTC m=+895.929268304" observedRunningTime="2026-03-11 13:37:29.033887821 +0000 UTC m=+896.715789545" watchObservedRunningTime="2026-03-11 13:37:29.03889829 +0000 UTC m=+896.720800044" Mar 11 13:37:29 crc kubenswrapper[4923]: I0311 13:37:29.056876 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-64jx7" podStartSLOduration=2.702133094 podStartE2EDuration="11.056859257s" podCreationTimestamp="2026-03-11 13:37:18 +0000 UTC" firstStartedPulling="2026-03-11 13:37:19.83887406 +0000 UTC m=+887.520775774" lastFinishedPulling="2026-03-11 13:37:28.193600223 +0000 UTC m=+895.875501937" observedRunningTime="2026-03-11 13:37:29.05483653 +0000 UTC m=+896.736738254" watchObservedRunningTime="2026-03-11 13:37:29.056859257 +0000 UTC m=+896.738760971" Mar 11 13:37:29 crc kubenswrapper[4923]: I0311 13:37:29.423844 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-64jx7" Mar 11 13:37:29 crc kubenswrapper[4923]: I0311 13:37:29.511658 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbg4d\" (UniqueName: \"kubernetes.io/projected/aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c-kube-api-access-cbg4d\") pod \"aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c\" (UID: \"aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c\") " Mar 11 13:37:29 crc kubenswrapper[4923]: I0311 13:37:29.516505 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c-kube-api-access-cbg4d" (OuterVolumeSpecName: "kube-api-access-cbg4d") pod "aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c" (UID: "aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c"). InnerVolumeSpecName "kube-api-access-cbg4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:37:29 crc kubenswrapper[4923]: I0311 13:37:29.613568 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbg4d\" (UniqueName: \"kubernetes.io/projected/aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c-kube-api-access-cbg4d\") on node \"crc\" DevicePath \"\"" Mar 11 13:37:30 crc kubenswrapper[4923]: I0311 13:37:30.025952 4923 generic.go:334] "Generic (PLEG): container finished" podID="aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c" containerID="43dfddabcd0b748bf7114df586ef20f07bb807379213eb47ba0114b5c88bfc6d" exitCode=0 Mar 11 13:37:30 crc kubenswrapper[4923]: I0311 13:37:30.026025 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-64jx7" Mar 11 13:37:30 crc kubenswrapper[4923]: I0311 13:37:30.026087 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-64jx7" event={"ID":"aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c","Type":"ContainerDied","Data":"43dfddabcd0b748bf7114df586ef20f07bb807379213eb47ba0114b5c88bfc6d"} Mar 11 13:37:30 crc kubenswrapper[4923]: I0311 13:37:30.026484 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-64jx7" event={"ID":"aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c","Type":"ContainerDied","Data":"151179b3514d2837b6914072e9eea3fc330257a711379cc1e624133f329beadb"} Mar 11 13:37:30 crc kubenswrapper[4923]: I0311 13:37:30.026522 4923 scope.go:117] "RemoveContainer" containerID="43dfddabcd0b748bf7114df586ef20f07bb807379213eb47ba0114b5c88bfc6d" Mar 11 13:37:30 crc kubenswrapper[4923]: I0311 13:37:30.028931 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-sd5l9" event={"ID":"9e845d8a-4bf4-457c-a8c3-871ed232846d","Type":"ContainerStarted","Data":"be126b0f53694072fe518465c2db2808b6701497c458dda8423ce0381a8c20df"} Mar 11 13:37:30 crc kubenswrapper[4923]: I0311 13:37:30.053608 4923 scope.go:117] "RemoveContainer" containerID="43dfddabcd0b748bf7114df586ef20f07bb807379213eb47ba0114b5c88bfc6d" Mar 11 13:37:30 crc kubenswrapper[4923]: E0311 13:37:30.054087 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43dfddabcd0b748bf7114df586ef20f07bb807379213eb47ba0114b5c88bfc6d\": container with ID starting with 43dfddabcd0b748bf7114df586ef20f07bb807379213eb47ba0114b5c88bfc6d not found: ID does not exist" containerID="43dfddabcd0b748bf7114df586ef20f07bb807379213eb47ba0114b5c88bfc6d" Mar 11 13:37:30 crc kubenswrapper[4923]: I0311 13:37:30.054120 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43dfddabcd0b748bf7114df586ef20f07bb807379213eb47ba0114b5c88bfc6d"} err="failed to get container status \"43dfddabcd0b748bf7114df586ef20f07bb807379213eb47ba0114b5c88bfc6d\": rpc error: code = NotFound desc = could not find container \"43dfddabcd0b748bf7114df586ef20f07bb807379213eb47ba0114b5c88bfc6d\": container with ID starting with 43dfddabcd0b748bf7114df586ef20f07bb807379213eb47ba0114b5c88bfc6d not found: ID does not exist" Mar 11 13:37:30 crc kubenswrapper[4923]: I0311 13:37:30.056658 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-sd5l9" podStartSLOduration=6.466557726 podStartE2EDuration="7.05664178s" podCreationTimestamp="2026-03-11 13:37:23 +0000 UTC" firstStartedPulling="2026-03-11 13:37:28.639653139 +0000 UTC m=+896.321554853" lastFinishedPulling="2026-03-11 13:37:29.229737173 +0000 UTC m=+896.911638907" observedRunningTime="2026-03-11 13:37:30.048194367 +0000 UTC m=+897.730096121" watchObservedRunningTime="2026-03-11 13:37:30.05664178 +0000 UTC m=+897.738543494" Mar 11 13:37:30 crc kubenswrapper[4923]: I0311 13:37:30.066048 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-64jx7"] Mar 11 13:37:30 crc kubenswrapper[4923]: I0311 13:37:30.070828 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-64jx7"] Mar 11 13:37:31 crc kubenswrapper[4923]: I0311 13:37:31.030168 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c" path="/var/lib/kubelet/pods/aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c/volumes" Mar 11 13:37:34 crc kubenswrapper[4923]: I0311 13:37:34.148219 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-sd5l9" Mar 11 13:37:34 crc kubenswrapper[4923]: I0311 13:37:34.149144 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-sd5l9" Mar 11 13:37:34 crc kubenswrapper[4923]: I0311 13:37:34.192788 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-sd5l9" Mar 11 13:37:35 crc kubenswrapper[4923]: I0311 13:37:35.106149 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-sd5l9" Mar 11 13:37:36 crc kubenswrapper[4923]: I0311 13:37:36.209533 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/memcached-0" Mar 11 13:37:36 crc kubenswrapper[4923]: I0311 13:37:36.830061 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94"] Mar 11 13:37:36 crc kubenswrapper[4923]: E0311 13:37:36.830285 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c" containerName="registry-server" Mar 11 13:37:36 crc kubenswrapper[4923]: I0311 13:37:36.830297 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c" containerName="registry-server" Mar 11 13:37:36 crc kubenswrapper[4923]: E0311 13:37:36.830303 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1da28d4d-80b8-4a02-8009-0b6032044448" containerName="mariadb-account-create-update" Mar 11 13:37:36 crc kubenswrapper[4923]: I0311 13:37:36.830310 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="1da28d4d-80b8-4a02-8009-0b6032044448" containerName="mariadb-account-create-update" Mar 11 13:37:36 crc kubenswrapper[4923]: I0311 13:37:36.830439 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="1da28d4d-80b8-4a02-8009-0b6032044448" containerName="mariadb-account-create-update" Mar 11 13:37:36 crc kubenswrapper[4923]: I0311 13:37:36.830455 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="aaad4f2b-bbc4-4f30-807a-c1f1a0ecaa3c" containerName="registry-server" Mar 11 13:37:36 crc kubenswrapper[4923]: I0311 13:37:36.831223 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94" Mar 11 13:37:36 crc kubenswrapper[4923]: I0311 13:37:36.833603 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-qkqtx" Mar 11 13:37:36 crc kubenswrapper[4923]: I0311 13:37:36.859546 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94"] Mar 11 13:37:36 crc kubenswrapper[4923]: I0311 13:37:36.916856 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/df8084e0-b75d-4c85-93ea-61cf2e3281b6-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94\" (UID: \"df8084e0-b75d-4c85-93ea-61cf2e3281b6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94" Mar 11 13:37:36 crc kubenswrapper[4923]: I0311 13:37:36.916921 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqmt4\" (UniqueName: \"kubernetes.io/projected/df8084e0-b75d-4c85-93ea-61cf2e3281b6-kube-api-access-jqmt4\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94\" (UID: \"df8084e0-b75d-4c85-93ea-61cf2e3281b6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94" Mar 11 13:37:36 crc kubenswrapper[4923]: I0311 13:37:36.916943 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/df8084e0-b75d-4c85-93ea-61cf2e3281b6-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94\" (UID: \"df8084e0-b75d-4c85-93ea-61cf2e3281b6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94" Mar 11 13:37:37 crc kubenswrapper[4923]: I0311 13:37:37.018804 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/df8084e0-b75d-4c85-93ea-61cf2e3281b6-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94\" (UID: \"df8084e0-b75d-4c85-93ea-61cf2e3281b6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94" Mar 11 13:37:37 crc kubenswrapper[4923]: I0311 13:37:37.019150 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqmt4\" (UniqueName: \"kubernetes.io/projected/df8084e0-b75d-4c85-93ea-61cf2e3281b6-kube-api-access-jqmt4\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94\" (UID: \"df8084e0-b75d-4c85-93ea-61cf2e3281b6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94" Mar 11 13:37:37 crc kubenswrapper[4923]: I0311 13:37:37.019267 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/df8084e0-b75d-4c85-93ea-61cf2e3281b6-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94\" (UID: \"df8084e0-b75d-4c85-93ea-61cf2e3281b6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94" Mar 11 13:37:37 crc kubenswrapper[4923]: I0311 13:37:37.019440 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/df8084e0-b75d-4c85-93ea-61cf2e3281b6-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94\" (UID: \"df8084e0-b75d-4c85-93ea-61cf2e3281b6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94" Mar 11 13:37:37 crc kubenswrapper[4923]: I0311 13:37:37.019745 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/df8084e0-b75d-4c85-93ea-61cf2e3281b6-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94\" (UID: \"df8084e0-b75d-4c85-93ea-61cf2e3281b6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94" Mar 11 13:37:37 crc kubenswrapper[4923]: I0311 13:37:37.044588 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqmt4\" (UniqueName: \"kubernetes.io/projected/df8084e0-b75d-4c85-93ea-61cf2e3281b6-kube-api-access-jqmt4\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94\" (UID: \"df8084e0-b75d-4c85-93ea-61cf2e3281b6\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94" Mar 11 13:37:37 crc kubenswrapper[4923]: I0311 13:37:37.151596 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94" Mar 11 13:37:37 crc kubenswrapper[4923]: I0311 13:37:37.574268 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94"] Mar 11 13:37:37 crc kubenswrapper[4923]: W0311 13:37:37.587021 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf8084e0_b75d_4c85_93ea_61cf2e3281b6.slice/crio-3e809a54a7b96805b4ddb2d0aff5cb54a4156b751a2b7a0b72573fc4edd11329 WatchSource:0}: Error finding container 3e809a54a7b96805b4ddb2d0aff5cb54a4156b751a2b7a0b72573fc4edd11329: Status 404 returned error can't find the container with id 3e809a54a7b96805b4ddb2d0aff5cb54a4156b751a2b7a0b72573fc4edd11329 Mar 11 13:37:38 crc kubenswrapper[4923]: I0311 13:37:38.103046 4923 generic.go:334] "Generic (PLEG): container finished" podID="df8084e0-b75d-4c85-93ea-61cf2e3281b6" containerID="b3d9f1d4d2b1a5e3b39bfd6c1128dcbeec65c1a5bd6116a6ab3352a13eb0704f" exitCode=0 Mar 11 13:37:38 crc kubenswrapper[4923]: I0311 13:37:38.103146 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94" event={"ID":"df8084e0-b75d-4c85-93ea-61cf2e3281b6","Type":"ContainerDied","Data":"b3d9f1d4d2b1a5e3b39bfd6c1128dcbeec65c1a5bd6116a6ab3352a13eb0704f"} Mar 11 13:37:38 crc kubenswrapper[4923]: I0311 13:37:38.103593 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94" event={"ID":"df8084e0-b75d-4c85-93ea-61cf2e3281b6","Type":"ContainerStarted","Data":"3e809a54a7b96805b4ddb2d0aff5cb54a4156b751a2b7a0b72573fc4edd11329"} Mar 11 13:37:40 crc kubenswrapper[4923]: I0311 13:37:40.121606 4923 generic.go:334] "Generic (PLEG): container finished" podID="df8084e0-b75d-4c85-93ea-61cf2e3281b6" containerID="6294f6c3d5f34c99ba8b66c063a582cb3f9b3e5f8af4d44c5ed25693ebe6cda5" exitCode=0 Mar 11 13:37:40 crc kubenswrapper[4923]: I0311 13:37:40.121663 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94" event={"ID":"df8084e0-b75d-4c85-93ea-61cf2e3281b6","Type":"ContainerDied","Data":"6294f6c3d5f34c99ba8b66c063a582cb3f9b3e5f8af4d44c5ed25693ebe6cda5"} Mar 11 13:37:41 crc kubenswrapper[4923]: I0311 13:37:41.136559 4923 generic.go:334] "Generic (PLEG): container finished" podID="df8084e0-b75d-4c85-93ea-61cf2e3281b6" containerID="782677559a92e2a1576aeb0779f89edb793d0953d229141ce1e2670698157c0f" exitCode=0 Mar 11 13:37:41 crc kubenswrapper[4923]: I0311 13:37:41.136654 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94" event={"ID":"df8084e0-b75d-4c85-93ea-61cf2e3281b6","Type":"ContainerDied","Data":"782677559a92e2a1576aeb0779f89edb793d0953d229141ce1e2670698157c0f"} Mar 11 13:37:41 crc kubenswrapper[4923]: I0311 13:37:41.931709 4923 patch_prober.go:28] interesting pod/machine-config-daemon-fsfnw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 11 13:37:41 crc kubenswrapper[4923]: I0311 13:37:41.931769 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 11 13:37:42 crc kubenswrapper[4923]: I0311 13:37:42.516415 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94" Mar 11 13:37:42 crc kubenswrapper[4923]: I0311 13:37:42.632930 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqmt4\" (UniqueName: \"kubernetes.io/projected/df8084e0-b75d-4c85-93ea-61cf2e3281b6-kube-api-access-jqmt4\") pod \"df8084e0-b75d-4c85-93ea-61cf2e3281b6\" (UID: \"df8084e0-b75d-4c85-93ea-61cf2e3281b6\") " Mar 11 13:37:42 crc kubenswrapper[4923]: I0311 13:37:42.633198 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/df8084e0-b75d-4c85-93ea-61cf2e3281b6-util\") pod \"df8084e0-b75d-4c85-93ea-61cf2e3281b6\" (UID: \"df8084e0-b75d-4c85-93ea-61cf2e3281b6\") " Mar 11 13:37:42 crc kubenswrapper[4923]: I0311 13:37:42.633276 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/df8084e0-b75d-4c85-93ea-61cf2e3281b6-bundle\") pod \"df8084e0-b75d-4c85-93ea-61cf2e3281b6\" (UID: \"df8084e0-b75d-4c85-93ea-61cf2e3281b6\") " Mar 11 13:37:42 crc kubenswrapper[4923]: I0311 13:37:42.634107 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df8084e0-b75d-4c85-93ea-61cf2e3281b6-bundle" (OuterVolumeSpecName: "bundle") pod "df8084e0-b75d-4c85-93ea-61cf2e3281b6" (UID: "df8084e0-b75d-4c85-93ea-61cf2e3281b6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:37:42 crc kubenswrapper[4923]: I0311 13:37:42.641183 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df8084e0-b75d-4c85-93ea-61cf2e3281b6-kube-api-access-jqmt4" (OuterVolumeSpecName: "kube-api-access-jqmt4") pod "df8084e0-b75d-4c85-93ea-61cf2e3281b6" (UID: "df8084e0-b75d-4c85-93ea-61cf2e3281b6"). InnerVolumeSpecName "kube-api-access-jqmt4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:37:42 crc kubenswrapper[4923]: I0311 13:37:42.659990 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df8084e0-b75d-4c85-93ea-61cf2e3281b6-util" (OuterVolumeSpecName: "util") pod "df8084e0-b75d-4c85-93ea-61cf2e3281b6" (UID: "df8084e0-b75d-4c85-93ea-61cf2e3281b6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:37:42 crc kubenswrapper[4923]: I0311 13:37:42.735229 4923 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/df8084e0-b75d-4c85-93ea-61cf2e3281b6-util\") on node \"crc\" DevicePath \"\"" Mar 11 13:37:42 crc kubenswrapper[4923]: I0311 13:37:42.735638 4923 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/df8084e0-b75d-4c85-93ea-61cf2e3281b6-bundle\") on node \"crc\" DevicePath \"\"" Mar 11 13:37:42 crc kubenswrapper[4923]: I0311 13:37:42.735776 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqmt4\" (UniqueName: \"kubernetes.io/projected/df8084e0-b75d-4c85-93ea-61cf2e3281b6-kube-api-access-jqmt4\") on node \"crc\" DevicePath \"\"" Mar 11 13:37:43 crc kubenswrapper[4923]: I0311 13:37:43.155024 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94" event={"ID":"df8084e0-b75d-4c85-93ea-61cf2e3281b6","Type":"ContainerDied","Data":"3e809a54a7b96805b4ddb2d0aff5cb54a4156b751a2b7a0b72573fc4edd11329"} Mar 11 13:37:43 crc kubenswrapper[4923]: I0311 13:37:43.155079 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94" Mar 11 13:37:43 crc kubenswrapper[4923]: I0311 13:37:43.155334 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e809a54a7b96805b4ddb2d0aff5cb54a4156b751a2b7a0b72573fc4edd11329" Mar 11 13:37:52 crc kubenswrapper[4923]: I0311 13:37:52.587237 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-5xpgn"] Mar 11 13:37:52 crc kubenswrapper[4923]: E0311 13:37:52.588090 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df8084e0-b75d-4c85-93ea-61cf2e3281b6" containerName="pull" Mar 11 13:37:52 crc kubenswrapper[4923]: I0311 13:37:52.588106 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="df8084e0-b75d-4c85-93ea-61cf2e3281b6" containerName="pull" Mar 11 13:37:52 crc kubenswrapper[4923]: E0311 13:37:52.588125 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df8084e0-b75d-4c85-93ea-61cf2e3281b6" containerName="util" Mar 11 13:37:52 crc kubenswrapper[4923]: I0311 13:37:52.588132 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="df8084e0-b75d-4c85-93ea-61cf2e3281b6" containerName="util" Mar 11 13:37:52 crc kubenswrapper[4923]: E0311 13:37:52.588149 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df8084e0-b75d-4c85-93ea-61cf2e3281b6" containerName="extract" Mar 11 13:37:52 crc kubenswrapper[4923]: I0311 13:37:52.588159 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="df8084e0-b75d-4c85-93ea-61cf2e3281b6" containerName="extract" Mar 11 13:37:52 crc kubenswrapper[4923]: I0311 13:37:52.588290 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="df8084e0-b75d-4c85-93ea-61cf2e3281b6" containerName="extract" Mar 11 13:37:52 crc kubenswrapper[4923]: I0311 13:37:52.588785 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-5xpgn" Mar 11 13:37:52 crc kubenswrapper[4923]: I0311 13:37:52.591476 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-tpqgl" Mar 11 13:37:52 crc kubenswrapper[4923]: I0311 13:37:52.615207 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-5xpgn"] Mar 11 13:37:52 crc kubenswrapper[4923]: I0311 13:37:52.680918 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zfm7\" (UniqueName: \"kubernetes.io/projected/eb048f52-cfe4-4827-9670-6406a758f39f-kube-api-access-4zfm7\") pod \"rabbitmq-cluster-operator-779fc9694b-5xpgn\" (UID: \"eb048f52-cfe4-4827-9670-6406a758f39f\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-5xpgn" Mar 11 13:37:52 crc kubenswrapper[4923]: I0311 13:37:52.782631 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zfm7\" (UniqueName: \"kubernetes.io/projected/eb048f52-cfe4-4827-9670-6406a758f39f-kube-api-access-4zfm7\") pod \"rabbitmq-cluster-operator-779fc9694b-5xpgn\" (UID: \"eb048f52-cfe4-4827-9670-6406a758f39f\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-5xpgn" Mar 11 13:37:52 crc kubenswrapper[4923]: I0311 13:37:52.805007 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zfm7\" (UniqueName: \"kubernetes.io/projected/eb048f52-cfe4-4827-9670-6406a758f39f-kube-api-access-4zfm7\") pod \"rabbitmq-cluster-operator-779fc9694b-5xpgn\" (UID: \"eb048f52-cfe4-4827-9670-6406a758f39f\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-5xpgn" Mar 11 13:37:52 crc kubenswrapper[4923]: I0311 13:37:52.911667 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-5xpgn" Mar 11 13:37:53 crc kubenswrapper[4923]: I0311 13:37:53.395671 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-5xpgn"] Mar 11 13:37:54 crc kubenswrapper[4923]: I0311 13:37:54.234483 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-5xpgn" event={"ID":"eb048f52-cfe4-4827-9670-6406a758f39f","Type":"ContainerStarted","Data":"16b201f8a46cb4b93aa59d1a58398dd4675b3a7b0c10ebc959eae3e3b08bca7e"} Mar 11 13:37:57 crc kubenswrapper[4923]: I0311 13:37:57.257110 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-5xpgn" event={"ID":"eb048f52-cfe4-4827-9670-6406a758f39f","Type":"ContainerStarted","Data":"766db4bf188d7b16622a87b840a7411c481d6db5384efddc2f2d657ab731a349"} Mar 11 13:37:57 crc kubenswrapper[4923]: I0311 13:37:57.283563 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-5xpgn" podStartSLOduration=1.95235797 podStartE2EDuration="5.283527789s" podCreationTimestamp="2026-03-11 13:37:52 +0000 UTC" firstStartedPulling="2026-03-11 13:37:53.400208686 +0000 UTC m=+921.082110400" lastFinishedPulling="2026-03-11 13:37:56.731378505 +0000 UTC m=+924.413280219" observedRunningTime="2026-03-11 13:37:57.277730008 +0000 UTC m=+924.959631802" watchObservedRunningTime="2026-03-11 13:37:57.283527789 +0000 UTC m=+924.965429553" Mar 11 13:38:00 crc kubenswrapper[4923]: I0311 13:38:00.143039 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29553938-tmvzr"] Mar 11 13:38:00 crc kubenswrapper[4923]: I0311 13:38:00.145149 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553938-tmvzr" Mar 11 13:38:00 crc kubenswrapper[4923]: I0311 13:38:00.147767 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 11 13:38:00 crc kubenswrapper[4923]: I0311 13:38:00.147843 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-nxck7" Mar 11 13:38:00 crc kubenswrapper[4923]: I0311 13:38:00.148938 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 11 13:38:00 crc kubenswrapper[4923]: I0311 13:38:00.168860 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553938-tmvzr"] Mar 11 13:38:00 crc kubenswrapper[4923]: I0311 13:38:00.291843 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwn8g\" (UniqueName: \"kubernetes.io/projected/91414f8a-2c8d-4aa0-8d82-3dd35377ef4c-kube-api-access-hwn8g\") pod \"auto-csr-approver-29553938-tmvzr\" (UID: \"91414f8a-2c8d-4aa0-8d82-3dd35377ef4c\") " pod="openshift-infra/auto-csr-approver-29553938-tmvzr" Mar 11 13:38:00 crc kubenswrapper[4923]: I0311 13:38:00.392900 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwn8g\" (UniqueName: \"kubernetes.io/projected/91414f8a-2c8d-4aa0-8d82-3dd35377ef4c-kube-api-access-hwn8g\") pod \"auto-csr-approver-29553938-tmvzr\" (UID: \"91414f8a-2c8d-4aa0-8d82-3dd35377ef4c\") " pod="openshift-infra/auto-csr-approver-29553938-tmvzr" Mar 11 13:38:00 crc kubenswrapper[4923]: I0311 13:38:00.413635 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwn8g\" (UniqueName: \"kubernetes.io/projected/91414f8a-2c8d-4aa0-8d82-3dd35377ef4c-kube-api-access-hwn8g\") pod \"auto-csr-approver-29553938-tmvzr\" (UID: \"91414f8a-2c8d-4aa0-8d82-3dd35377ef4c\") " pod="openshift-infra/auto-csr-approver-29553938-tmvzr" Mar 11 13:38:00 crc kubenswrapper[4923]: I0311 13:38:00.470483 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553938-tmvzr" Mar 11 13:38:00 crc kubenswrapper[4923]: I0311 13:38:00.731375 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553938-tmvzr"] Mar 11 13:38:01 crc kubenswrapper[4923]: I0311 13:38:01.295822 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553938-tmvzr" event={"ID":"91414f8a-2c8d-4aa0-8d82-3dd35377ef4c","Type":"ContainerStarted","Data":"e79bc2d6b80bc6f6479ca0d9c17d4541de3d53c56f25e3069891cfcc3062c5ce"} Mar 11 13:38:02 crc kubenswrapper[4923]: I0311 13:38:02.304558 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553938-tmvzr" event={"ID":"91414f8a-2c8d-4aa0-8d82-3dd35377ef4c","Type":"ContainerStarted","Data":"f53ae113b8c07d010ea14334e7eb25e74c34712b50222a8106bbce8973c62c8a"} Mar 11 13:38:02 crc kubenswrapper[4923]: I0311 13:38:02.456578 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29553938-tmvzr" podStartSLOduration=1.3044933 podStartE2EDuration="2.4565429s" podCreationTimestamp="2026-03-11 13:38:00 +0000 UTC" firstStartedPulling="2026-03-11 13:38:00.735049342 +0000 UTC m=+928.416951066" lastFinishedPulling="2026-03-11 13:38:01.887098952 +0000 UTC m=+929.569000666" observedRunningTime="2026-03-11 13:38:02.445395213 +0000 UTC m=+930.127296927" watchObservedRunningTime="2026-03-11 13:38:02.4565429 +0000 UTC m=+930.138444654" Mar 11 13:38:03 crc kubenswrapper[4923]: I0311 13:38:03.314125 4923 generic.go:334] "Generic (PLEG): container finished" podID="91414f8a-2c8d-4aa0-8d82-3dd35377ef4c" containerID="f53ae113b8c07d010ea14334e7eb25e74c34712b50222a8106bbce8973c62c8a" exitCode=0 Mar 11 13:38:03 crc kubenswrapper[4923]: I0311 13:38:03.314171 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553938-tmvzr" event={"ID":"91414f8a-2c8d-4aa0-8d82-3dd35377ef4c","Type":"ContainerDied","Data":"f53ae113b8c07d010ea14334e7eb25e74c34712b50222a8106bbce8973c62c8a"} Mar 11 13:38:03 crc kubenswrapper[4923]: I0311 13:38:03.761560 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Mar 11 13:38:03 crc kubenswrapper[4923]: I0311 13:38:03.762494 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:03 crc kubenswrapper[4923]: I0311 13:38:03.764625 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"rabbitmq-default-user" Mar 11 13:38:03 crc kubenswrapper[4923]: I0311 13:38:03.764489 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"rabbitmq-erlang-cookie" Mar 11 13:38:03 crc kubenswrapper[4923]: I0311 13:38:03.765728 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"rabbitmq-plugins-conf" Mar 11 13:38:03 crc kubenswrapper[4923]: I0311 13:38:03.765907 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"rabbitmq-server-dockercfg-9xkqc" Mar 11 13:38:03 crc kubenswrapper[4923]: I0311 13:38:03.767920 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cinder-kuttl-tests"/"rabbitmq-server-conf" Mar 11 13:38:03 crc kubenswrapper[4923]: I0311 13:38:03.775593 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Mar 11 13:38:03 crc kubenswrapper[4923]: I0311 13:38:03.922015 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/50119944-78f3-4b19-b19b-1bc87e93e3fc-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:03 crc kubenswrapper[4923]: I0311 13:38:03.922166 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/50119944-78f3-4b19-b19b-1bc87e93e3fc-pod-info\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:03 crc kubenswrapper[4923]: I0311 13:38:03.922284 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/50119944-78f3-4b19-b19b-1bc87e93e3fc-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:03 crc kubenswrapper[4923]: I0311 13:38:03.922363 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ec7102e8-ec30-4d5c-abaa-e946b770c548\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ec7102e8-ec30-4d5c-abaa-e946b770c548\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:03 crc kubenswrapper[4923]: I0311 13:38:03.922389 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/50119944-78f3-4b19-b19b-1bc87e93e3fc-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:03 crc kubenswrapper[4923]: I0311 13:38:03.922616 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lghh6\" (UniqueName: \"kubernetes.io/projected/50119944-78f3-4b19-b19b-1bc87e93e3fc-kube-api-access-lghh6\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:03 crc kubenswrapper[4923]: I0311 13:38:03.922689 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/50119944-78f3-4b19-b19b-1bc87e93e3fc-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:03 crc kubenswrapper[4923]: I0311 13:38:03.922718 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/50119944-78f3-4b19-b19b-1bc87e93e3fc-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.023990 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/50119944-78f3-4b19-b19b-1bc87e93e3fc-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.024057 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/50119944-78f3-4b19-b19b-1bc87e93e3fc-pod-info\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.024127 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/50119944-78f3-4b19-b19b-1bc87e93e3fc-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.024151 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lghh6\" (UniqueName: \"kubernetes.io/projected/50119944-78f3-4b19-b19b-1bc87e93e3fc-kube-api-access-lghh6\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.024174 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ec7102e8-ec30-4d5c-abaa-e946b770c548\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ec7102e8-ec30-4d5c-abaa-e946b770c548\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.024196 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/50119944-78f3-4b19-b19b-1bc87e93e3fc-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.024223 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/50119944-78f3-4b19-b19b-1bc87e93e3fc-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.024250 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/50119944-78f3-4b19-b19b-1bc87e93e3fc-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.025022 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/50119944-78f3-4b19-b19b-1bc87e93e3fc-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.026115 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/50119944-78f3-4b19-b19b-1bc87e93e3fc-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.026502 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/50119944-78f3-4b19-b19b-1bc87e93e3fc-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.028184 4923 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.028221 4923 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ec7102e8-ec30-4d5c-abaa-e946b770c548\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ec7102e8-ec30-4d5c-abaa-e946b770c548\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/95ca7890bc92da3513cb525a5e7fe14c9669fcc4516444131ea09c56ff34f53c/globalmount\"" pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.031333 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/50119944-78f3-4b19-b19b-1bc87e93e3fc-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.031827 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/50119944-78f3-4b19-b19b-1bc87e93e3fc-pod-info\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.033049 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/50119944-78f3-4b19-b19b-1bc87e93e3fc-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.041468 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lghh6\" (UniqueName: \"kubernetes.io/projected/50119944-78f3-4b19-b19b-1bc87e93e3fc-kube-api-access-lghh6\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.054980 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ec7102e8-ec30-4d5c-abaa-e946b770c548\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ec7102e8-ec30-4d5c-abaa-e946b770c548\") pod \"rabbitmq-server-0\" (UID: \"50119944-78f3-4b19-b19b-1bc87e93e3fc\") " pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.087010 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.586948 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553938-tmvzr" Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.627201 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/rabbitmq-server-0"] Mar 11 13:38:04 crc kubenswrapper[4923]: W0311 13:38:04.637861 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50119944_78f3_4b19_b19b_1bc87e93e3fc.slice/crio-fd318d1d624b5bbb32dec9792e0eaf4372980d632ab3e2a0458b166e6588e6a8 WatchSource:0}: Error finding container fd318d1d624b5bbb32dec9792e0eaf4372980d632ab3e2a0458b166e6588e6a8: Status 404 returned error can't find the container with id fd318d1d624b5bbb32dec9792e0eaf4372980d632ab3e2a0458b166e6588e6a8 Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.647956 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwn8g\" (UniqueName: \"kubernetes.io/projected/91414f8a-2c8d-4aa0-8d82-3dd35377ef4c-kube-api-access-hwn8g\") pod \"91414f8a-2c8d-4aa0-8d82-3dd35377ef4c\" (UID: \"91414f8a-2c8d-4aa0-8d82-3dd35377ef4c\") " Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.655646 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91414f8a-2c8d-4aa0-8d82-3dd35377ef4c-kube-api-access-hwn8g" (OuterVolumeSpecName: "kube-api-access-hwn8g") pod "91414f8a-2c8d-4aa0-8d82-3dd35377ef4c" (UID: "91414f8a-2c8d-4aa0-8d82-3dd35377ef4c"). InnerVolumeSpecName "kube-api-access-hwn8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:38:04 crc kubenswrapper[4923]: I0311 13:38:04.750210 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwn8g\" (UniqueName: \"kubernetes.io/projected/91414f8a-2c8d-4aa0-8d82-3dd35377ef4c-kube-api-access-hwn8g\") on node \"crc\" DevicePath \"\"" Mar 11 13:38:05 crc kubenswrapper[4923]: I0311 13:38:05.357950 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553938-tmvzr" event={"ID":"91414f8a-2c8d-4aa0-8d82-3dd35377ef4c","Type":"ContainerDied","Data":"e79bc2d6b80bc6f6479ca0d9c17d4541de3d53c56f25e3069891cfcc3062c5ce"} Mar 11 13:38:05 crc kubenswrapper[4923]: I0311 13:38:05.357999 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e79bc2d6b80bc6f6479ca0d9c17d4541de3d53c56f25e3069891cfcc3062c5ce" Mar 11 13:38:05 crc kubenswrapper[4923]: I0311 13:38:05.358079 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553938-tmvzr" Mar 11 13:38:05 crc kubenswrapper[4923]: I0311 13:38:05.359902 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"50119944-78f3-4b19-b19b-1bc87e93e3fc","Type":"ContainerStarted","Data":"fd318d1d624b5bbb32dec9792e0eaf4372980d632ab3e2a0458b166e6588e6a8"} Mar 11 13:38:05 crc kubenswrapper[4923]: I0311 13:38:05.399438 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-5jw2x"] Mar 11 13:38:05 crc kubenswrapper[4923]: E0311 13:38:05.399974 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91414f8a-2c8d-4aa0-8d82-3dd35377ef4c" containerName="oc" Mar 11 13:38:05 crc kubenswrapper[4923]: I0311 13:38:05.400008 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="91414f8a-2c8d-4aa0-8d82-3dd35377ef4c" containerName="oc" Mar 11 13:38:05 crc kubenswrapper[4923]: I0311 13:38:05.400214 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="91414f8a-2c8d-4aa0-8d82-3dd35377ef4c" containerName="oc" Mar 11 13:38:05 crc kubenswrapper[4923]: I0311 13:38:05.401069 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-5jw2x" Mar 11 13:38:05 crc kubenswrapper[4923]: I0311 13:38:05.402885 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-krwdn" Mar 11 13:38:05 crc kubenswrapper[4923]: I0311 13:38:05.404227 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-5jw2x"] Mar 11 13:38:05 crc kubenswrapper[4923]: I0311 13:38:05.461306 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48sct\" (UniqueName: \"kubernetes.io/projected/ff2988d3-36a1-4085-935a-6d9ab916faac-kube-api-access-48sct\") pod \"keystone-operator-index-5jw2x\" (UID: \"ff2988d3-36a1-4085-935a-6d9ab916faac\") " pod="openstack-operators/keystone-operator-index-5jw2x" Mar 11 13:38:05 crc kubenswrapper[4923]: I0311 13:38:05.563037 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48sct\" (UniqueName: \"kubernetes.io/projected/ff2988d3-36a1-4085-935a-6d9ab916faac-kube-api-access-48sct\") pod \"keystone-operator-index-5jw2x\" (UID: \"ff2988d3-36a1-4085-935a-6d9ab916faac\") " pod="openstack-operators/keystone-operator-index-5jw2x" Mar 11 13:38:05 crc kubenswrapper[4923]: I0311 13:38:05.592905 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48sct\" (UniqueName: \"kubernetes.io/projected/ff2988d3-36a1-4085-935a-6d9ab916faac-kube-api-access-48sct\") pod \"keystone-operator-index-5jw2x\" (UID: \"ff2988d3-36a1-4085-935a-6d9ab916faac\") " pod="openstack-operators/keystone-operator-index-5jw2x" Mar 11 13:38:05 crc kubenswrapper[4923]: I0311 13:38:05.654661 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29553932-577nf"] Mar 11 13:38:05 crc kubenswrapper[4923]: I0311 13:38:05.661089 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29553932-577nf"] Mar 11 13:38:05 crc kubenswrapper[4923]: I0311 13:38:05.752349 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-5jw2x" Mar 11 13:38:06 crc kubenswrapper[4923]: I0311 13:38:06.172937 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-5jw2x"] Mar 11 13:38:06 crc kubenswrapper[4923]: W0311 13:38:06.188364 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podff2988d3_36a1_4085_935a_6d9ab916faac.slice/crio-54a7e52734eb77fccdc03a2db6046b29b691086ebbd36476c757dcaf88b04ab1 WatchSource:0}: Error finding container 54a7e52734eb77fccdc03a2db6046b29b691086ebbd36476c757dcaf88b04ab1: Status 404 returned error can't find the container with id 54a7e52734eb77fccdc03a2db6046b29b691086ebbd36476c757dcaf88b04ab1 Mar 11 13:38:06 crc kubenswrapper[4923]: I0311 13:38:06.367524 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-5jw2x" event={"ID":"ff2988d3-36a1-4085-935a-6d9ab916faac","Type":"ContainerStarted","Data":"54a7e52734eb77fccdc03a2db6046b29b691086ebbd36476c757dcaf88b04ab1"} Mar 11 13:38:07 crc kubenswrapper[4923]: I0311 13:38:07.038730 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39d4c595-2048-4b75-9c9e-e7d7d817afd7" path="/var/lib/kubelet/pods/39d4c595-2048-4b75-9c9e-e7d7d817afd7/volumes" Mar 11 13:38:11 crc kubenswrapper[4923]: I0311 13:38:11.418825 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-5jw2x" event={"ID":"ff2988d3-36a1-4085-935a-6d9ab916faac","Type":"ContainerStarted","Data":"bf9006966ff874007b3a4e8e2f3dbb94fa27a3d7bcb6fe5266f33cc84863a0da"} Mar 11 13:38:11 crc kubenswrapper[4923]: I0311 13:38:11.446416 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-5jw2x" podStartSLOduration=1.646476645 podStartE2EDuration="6.446393463s" podCreationTimestamp="2026-03-11 13:38:05 +0000 UTC" firstStartedPulling="2026-03-11 13:38:06.192380942 +0000 UTC m=+933.874282656" lastFinishedPulling="2026-03-11 13:38:10.99229774 +0000 UTC m=+938.674199474" observedRunningTime="2026-03-11 13:38:11.443091682 +0000 UTC m=+939.124993406" watchObservedRunningTime="2026-03-11 13:38:11.446393463 +0000 UTC m=+939.128295187" Mar 11 13:38:11 crc kubenswrapper[4923]: I0311 13:38:11.932469 4923 patch_prober.go:28] interesting pod/machine-config-daemon-fsfnw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 11 13:38:11 crc kubenswrapper[4923]: I0311 13:38:11.932552 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 11 13:38:11 crc kubenswrapper[4923]: I0311 13:38:11.932612 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" Mar 11 13:38:11 crc kubenswrapper[4923]: I0311 13:38:11.933420 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a555ce4420a87a2fc08d149d6004aea3fd8a76f007c748edd82c76ca49ddb60a"} pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 11 13:38:11 crc kubenswrapper[4923]: I0311 13:38:11.933508 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" containerID="cri-o://a555ce4420a87a2fc08d149d6004aea3fd8a76f007c748edd82c76ca49ddb60a" gracePeriod=600 Mar 11 13:38:12 crc kubenswrapper[4923]: I0311 13:38:12.428404 4923 generic.go:334] "Generic (PLEG): container finished" podID="356de301-ded3-4112-8406-1a28e44c2082" containerID="a555ce4420a87a2fc08d149d6004aea3fd8a76f007c748edd82c76ca49ddb60a" exitCode=0 Mar 11 13:38:12 crc kubenswrapper[4923]: I0311 13:38:12.428552 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" event={"ID":"356de301-ded3-4112-8406-1a28e44c2082","Type":"ContainerDied","Data":"a555ce4420a87a2fc08d149d6004aea3fd8a76f007c748edd82c76ca49ddb60a"} Mar 11 13:38:12 crc kubenswrapper[4923]: I0311 13:38:12.428796 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" event={"ID":"356de301-ded3-4112-8406-1a28e44c2082","Type":"ContainerStarted","Data":"ead0c2f701ab9961c880f0090716c64783b2bee2562b04e2163c46d120ab88f4"} Mar 11 13:38:12 crc kubenswrapper[4923]: I0311 13:38:12.428818 4923 scope.go:117] "RemoveContainer" containerID="e9d14dc9fced1f5ac360fb70b07bfe3d2a670f2c82a614a9b6443f96729aaa50" Mar 11 13:38:12 crc kubenswrapper[4923]: I0311 13:38:12.431234 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"50119944-78f3-4b19-b19b-1bc87e93e3fc","Type":"ContainerStarted","Data":"a4d278c08d19fb41723a97321fa1ccb51ebe02c480e860f8ff4dd9b8e2b1c166"} Mar 11 13:38:15 crc kubenswrapper[4923]: I0311 13:38:15.752571 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-5jw2x" Mar 11 13:38:15 crc kubenswrapper[4923]: I0311 13:38:15.754497 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-5jw2x" Mar 11 13:38:15 crc kubenswrapper[4923]: I0311 13:38:15.800756 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-5jw2x" Mar 11 13:38:16 crc kubenswrapper[4923]: I0311 13:38:16.497368 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-5jw2x" Mar 11 13:38:18 crc kubenswrapper[4923]: I0311 13:38:18.452154 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr"] Mar 11 13:38:18 crc kubenswrapper[4923]: I0311 13:38:18.454045 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr" Mar 11 13:38:18 crc kubenswrapper[4923]: I0311 13:38:18.456900 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-qkqtx" Mar 11 13:38:18 crc kubenswrapper[4923]: I0311 13:38:18.472575 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr"] Mar 11 13:38:18 crc kubenswrapper[4923]: I0311 13:38:18.549362 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fc311cce-a1ef-45e8-be23-2b5c55bb4243-util\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr\" (UID: \"fc311cce-a1ef-45e8-be23-2b5c55bb4243\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr" Mar 11 13:38:18 crc kubenswrapper[4923]: I0311 13:38:18.549502 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fc311cce-a1ef-45e8-be23-2b5c55bb4243-bundle\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr\" (UID: \"fc311cce-a1ef-45e8-be23-2b5c55bb4243\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr" Mar 11 13:38:18 crc kubenswrapper[4923]: I0311 13:38:18.550616 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vwctx\" (UniqueName: \"kubernetes.io/projected/fc311cce-a1ef-45e8-be23-2b5c55bb4243-kube-api-access-vwctx\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr\" (UID: \"fc311cce-a1ef-45e8-be23-2b5c55bb4243\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr" Mar 11 13:38:18 crc kubenswrapper[4923]: I0311 13:38:18.652218 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fc311cce-a1ef-45e8-be23-2b5c55bb4243-util\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr\" (UID: \"fc311cce-a1ef-45e8-be23-2b5c55bb4243\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr" Mar 11 13:38:18 crc kubenswrapper[4923]: I0311 13:38:18.652390 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fc311cce-a1ef-45e8-be23-2b5c55bb4243-bundle\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr\" (UID: \"fc311cce-a1ef-45e8-be23-2b5c55bb4243\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr" Mar 11 13:38:18 crc kubenswrapper[4923]: I0311 13:38:18.652438 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vwctx\" (UniqueName: \"kubernetes.io/projected/fc311cce-a1ef-45e8-be23-2b5c55bb4243-kube-api-access-vwctx\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr\" (UID: \"fc311cce-a1ef-45e8-be23-2b5c55bb4243\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr" Mar 11 13:38:18 crc kubenswrapper[4923]: I0311 13:38:18.652846 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fc311cce-a1ef-45e8-be23-2b5c55bb4243-util\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr\" (UID: \"fc311cce-a1ef-45e8-be23-2b5c55bb4243\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr" Mar 11 13:38:18 crc kubenswrapper[4923]: I0311 13:38:18.653651 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fc311cce-a1ef-45e8-be23-2b5c55bb4243-bundle\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr\" (UID: \"fc311cce-a1ef-45e8-be23-2b5c55bb4243\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr" Mar 11 13:38:18 crc kubenswrapper[4923]: I0311 13:38:18.690397 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vwctx\" (UniqueName: \"kubernetes.io/projected/fc311cce-a1ef-45e8-be23-2b5c55bb4243-kube-api-access-vwctx\") pod \"a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr\" (UID: \"fc311cce-a1ef-45e8-be23-2b5c55bb4243\") " pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr" Mar 11 13:38:18 crc kubenswrapper[4923]: I0311 13:38:18.834461 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr" Mar 11 13:38:19 crc kubenswrapper[4923]: I0311 13:38:19.299450 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr"] Mar 11 13:38:19 crc kubenswrapper[4923]: I0311 13:38:19.498519 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr" event={"ID":"fc311cce-a1ef-45e8-be23-2b5c55bb4243","Type":"ContainerStarted","Data":"d511cee209f04dea586ee576f1694e232ff4e4472a865b1d25314dafdd42a8c0"} Mar 11 13:38:19 crc kubenswrapper[4923]: I0311 13:38:19.498896 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr" event={"ID":"fc311cce-a1ef-45e8-be23-2b5c55bb4243","Type":"ContainerStarted","Data":"39984e7df5c773ee6ff998e387bc1ad303b50d9c03ca748de9827b6629efecdf"} Mar 11 13:38:20 crc kubenswrapper[4923]: I0311 13:38:20.509538 4923 generic.go:334] "Generic (PLEG): container finished" podID="fc311cce-a1ef-45e8-be23-2b5c55bb4243" containerID="d511cee209f04dea586ee576f1694e232ff4e4472a865b1d25314dafdd42a8c0" exitCode=0 Mar 11 13:38:20 crc kubenswrapper[4923]: I0311 13:38:20.509605 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr" event={"ID":"fc311cce-a1ef-45e8-be23-2b5c55bb4243","Type":"ContainerDied","Data":"d511cee209f04dea586ee576f1694e232ff4e4472a865b1d25314dafdd42a8c0"} Mar 11 13:38:22 crc kubenswrapper[4923]: I0311 13:38:22.536612 4923 generic.go:334] "Generic (PLEG): container finished" podID="fc311cce-a1ef-45e8-be23-2b5c55bb4243" containerID="e7d1ef45628a3bac9f9a87e1acc9c61d064d9ff0f5d3a99bdd3485ffb3180683" exitCode=0 Mar 11 13:38:22 crc kubenswrapper[4923]: I0311 13:38:22.536700 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr" event={"ID":"fc311cce-a1ef-45e8-be23-2b5c55bb4243","Type":"ContainerDied","Data":"e7d1ef45628a3bac9f9a87e1acc9c61d064d9ff0f5d3a99bdd3485ffb3180683"} Mar 11 13:38:23 crc kubenswrapper[4923]: I0311 13:38:23.544840 4923 generic.go:334] "Generic (PLEG): container finished" podID="fc311cce-a1ef-45e8-be23-2b5c55bb4243" containerID="39fac530904e0f536ab175aad8ed28ea2a21c110c1ca6f31a90eee98a99fa071" exitCode=0 Mar 11 13:38:23 crc kubenswrapper[4923]: I0311 13:38:23.544875 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr" event={"ID":"fc311cce-a1ef-45e8-be23-2b5c55bb4243","Type":"ContainerDied","Data":"39fac530904e0f536ab175aad8ed28ea2a21c110c1ca6f31a90eee98a99fa071"} Mar 11 13:38:24 crc kubenswrapper[4923]: I0311 13:38:24.886442 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr" Mar 11 13:38:25 crc kubenswrapper[4923]: I0311 13:38:25.052143 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vwctx\" (UniqueName: \"kubernetes.io/projected/fc311cce-a1ef-45e8-be23-2b5c55bb4243-kube-api-access-vwctx\") pod \"fc311cce-a1ef-45e8-be23-2b5c55bb4243\" (UID: \"fc311cce-a1ef-45e8-be23-2b5c55bb4243\") " Mar 11 13:38:25 crc kubenswrapper[4923]: I0311 13:38:25.052223 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fc311cce-a1ef-45e8-be23-2b5c55bb4243-bundle\") pod \"fc311cce-a1ef-45e8-be23-2b5c55bb4243\" (UID: \"fc311cce-a1ef-45e8-be23-2b5c55bb4243\") " Mar 11 13:38:25 crc kubenswrapper[4923]: I0311 13:38:25.052402 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fc311cce-a1ef-45e8-be23-2b5c55bb4243-util\") pod \"fc311cce-a1ef-45e8-be23-2b5c55bb4243\" (UID: \"fc311cce-a1ef-45e8-be23-2b5c55bb4243\") " Mar 11 13:38:25 crc kubenswrapper[4923]: I0311 13:38:25.053898 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc311cce-a1ef-45e8-be23-2b5c55bb4243-bundle" (OuterVolumeSpecName: "bundle") pod "fc311cce-a1ef-45e8-be23-2b5c55bb4243" (UID: "fc311cce-a1ef-45e8-be23-2b5c55bb4243"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:38:25 crc kubenswrapper[4923]: I0311 13:38:25.054175 4923 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fc311cce-a1ef-45e8-be23-2b5c55bb4243-bundle\") on node \"crc\" DevicePath \"\"" Mar 11 13:38:25 crc kubenswrapper[4923]: I0311 13:38:25.057621 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc311cce-a1ef-45e8-be23-2b5c55bb4243-kube-api-access-vwctx" (OuterVolumeSpecName: "kube-api-access-vwctx") pod "fc311cce-a1ef-45e8-be23-2b5c55bb4243" (UID: "fc311cce-a1ef-45e8-be23-2b5c55bb4243"). InnerVolumeSpecName "kube-api-access-vwctx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:38:25 crc kubenswrapper[4923]: I0311 13:38:25.068126 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc311cce-a1ef-45e8-be23-2b5c55bb4243-util" (OuterVolumeSpecName: "util") pod "fc311cce-a1ef-45e8-be23-2b5c55bb4243" (UID: "fc311cce-a1ef-45e8-be23-2b5c55bb4243"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:38:25 crc kubenswrapper[4923]: I0311 13:38:25.156391 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vwctx\" (UniqueName: \"kubernetes.io/projected/fc311cce-a1ef-45e8-be23-2b5c55bb4243-kube-api-access-vwctx\") on node \"crc\" DevicePath \"\"" Mar 11 13:38:25 crc kubenswrapper[4923]: I0311 13:38:25.156445 4923 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fc311cce-a1ef-45e8-be23-2b5c55bb4243-util\") on node \"crc\" DevicePath \"\"" Mar 11 13:38:25 crc kubenswrapper[4923]: I0311 13:38:25.561938 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr" event={"ID":"fc311cce-a1ef-45e8-be23-2b5c55bb4243","Type":"ContainerDied","Data":"39984e7df5c773ee6ff998e387bc1ad303b50d9c03ca748de9827b6629efecdf"} Mar 11 13:38:25 crc kubenswrapper[4923]: I0311 13:38:25.561992 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39984e7df5c773ee6ff998e387bc1ad303b50d9c03ca748de9827b6629efecdf" Mar 11 13:38:25 crc kubenswrapper[4923]: I0311 13:38:25.561997 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr" Mar 11 13:38:25 crc kubenswrapper[4923]: E0311 13:38:25.709883 4923 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc311cce_a1ef_45e8_be23_2b5c55bb4243.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc311cce_a1ef_45e8_be23_2b5c55bb4243.slice/crio-39984e7df5c773ee6ff998e387bc1ad303b50d9c03ca748de9827b6629efecdf\": RecentStats: unable to find data in memory cache]" Mar 11 13:38:33 crc kubenswrapper[4923]: I0311 13:38:33.305234 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7f7964dd58-bjc24"] Mar 11 13:38:33 crc kubenswrapper[4923]: E0311 13:38:33.305938 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc311cce-a1ef-45e8-be23-2b5c55bb4243" containerName="util" Mar 11 13:38:33 crc kubenswrapper[4923]: I0311 13:38:33.305952 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc311cce-a1ef-45e8-be23-2b5c55bb4243" containerName="util" Mar 11 13:38:33 crc kubenswrapper[4923]: E0311 13:38:33.305972 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc311cce-a1ef-45e8-be23-2b5c55bb4243" containerName="extract" Mar 11 13:38:33 crc kubenswrapper[4923]: I0311 13:38:33.305979 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc311cce-a1ef-45e8-be23-2b5c55bb4243" containerName="extract" Mar 11 13:38:33 crc kubenswrapper[4923]: E0311 13:38:33.305995 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc311cce-a1ef-45e8-be23-2b5c55bb4243" containerName="pull" Mar 11 13:38:33 crc kubenswrapper[4923]: I0311 13:38:33.306003 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc311cce-a1ef-45e8-be23-2b5c55bb4243" containerName="pull" Mar 11 13:38:33 crc kubenswrapper[4923]: I0311 13:38:33.306138 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc311cce-a1ef-45e8-be23-2b5c55bb4243" containerName="extract" Mar 11 13:38:33 crc kubenswrapper[4923]: I0311 13:38:33.306625 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7f7964dd58-bjc24" Mar 11 13:38:33 crc kubenswrapper[4923]: I0311 13:38:33.308757 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-lk2zf" Mar 11 13:38:33 crc kubenswrapper[4923]: I0311 13:38:33.308943 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Mar 11 13:38:33 crc kubenswrapper[4923]: I0311 13:38:33.326383 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7f7964dd58-bjc24"] Mar 11 13:38:33 crc kubenswrapper[4923]: I0311 13:38:33.377745 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a492ddb6-a936-4f7d-ba13-d4ddb83343a0-webhook-cert\") pod \"keystone-operator-controller-manager-7f7964dd58-bjc24\" (UID: \"a492ddb6-a936-4f7d-ba13-d4ddb83343a0\") " pod="openstack-operators/keystone-operator-controller-manager-7f7964dd58-bjc24" Mar 11 13:38:33 crc kubenswrapper[4923]: I0311 13:38:33.377822 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a492ddb6-a936-4f7d-ba13-d4ddb83343a0-apiservice-cert\") pod \"keystone-operator-controller-manager-7f7964dd58-bjc24\" (UID: \"a492ddb6-a936-4f7d-ba13-d4ddb83343a0\") " pod="openstack-operators/keystone-operator-controller-manager-7f7964dd58-bjc24" Mar 11 13:38:33 crc kubenswrapper[4923]: I0311 13:38:33.377856 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnb2x\" (UniqueName: \"kubernetes.io/projected/a492ddb6-a936-4f7d-ba13-d4ddb83343a0-kube-api-access-cnb2x\") pod \"keystone-operator-controller-manager-7f7964dd58-bjc24\" (UID: \"a492ddb6-a936-4f7d-ba13-d4ddb83343a0\") " pod="openstack-operators/keystone-operator-controller-manager-7f7964dd58-bjc24" Mar 11 13:38:33 crc kubenswrapper[4923]: I0311 13:38:33.479462 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a492ddb6-a936-4f7d-ba13-d4ddb83343a0-apiservice-cert\") pod \"keystone-operator-controller-manager-7f7964dd58-bjc24\" (UID: \"a492ddb6-a936-4f7d-ba13-d4ddb83343a0\") " pod="openstack-operators/keystone-operator-controller-manager-7f7964dd58-bjc24" Mar 11 13:38:33 crc kubenswrapper[4923]: I0311 13:38:33.479506 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnb2x\" (UniqueName: \"kubernetes.io/projected/a492ddb6-a936-4f7d-ba13-d4ddb83343a0-kube-api-access-cnb2x\") pod \"keystone-operator-controller-manager-7f7964dd58-bjc24\" (UID: \"a492ddb6-a936-4f7d-ba13-d4ddb83343a0\") " pod="openstack-operators/keystone-operator-controller-manager-7f7964dd58-bjc24" Mar 11 13:38:33 crc kubenswrapper[4923]: I0311 13:38:33.479585 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a492ddb6-a936-4f7d-ba13-d4ddb83343a0-webhook-cert\") pod \"keystone-operator-controller-manager-7f7964dd58-bjc24\" (UID: \"a492ddb6-a936-4f7d-ba13-d4ddb83343a0\") " pod="openstack-operators/keystone-operator-controller-manager-7f7964dd58-bjc24" Mar 11 13:38:33 crc kubenswrapper[4923]: I0311 13:38:33.486169 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a492ddb6-a936-4f7d-ba13-d4ddb83343a0-webhook-cert\") pod \"keystone-operator-controller-manager-7f7964dd58-bjc24\" (UID: \"a492ddb6-a936-4f7d-ba13-d4ddb83343a0\") " pod="openstack-operators/keystone-operator-controller-manager-7f7964dd58-bjc24" Mar 11 13:38:33 crc kubenswrapper[4923]: I0311 13:38:33.495011 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a492ddb6-a936-4f7d-ba13-d4ddb83343a0-apiservice-cert\") pod \"keystone-operator-controller-manager-7f7964dd58-bjc24\" (UID: \"a492ddb6-a936-4f7d-ba13-d4ddb83343a0\") " pod="openstack-operators/keystone-operator-controller-manager-7f7964dd58-bjc24" Mar 11 13:38:33 crc kubenswrapper[4923]: I0311 13:38:33.502103 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnb2x\" (UniqueName: \"kubernetes.io/projected/a492ddb6-a936-4f7d-ba13-d4ddb83343a0-kube-api-access-cnb2x\") pod \"keystone-operator-controller-manager-7f7964dd58-bjc24\" (UID: \"a492ddb6-a936-4f7d-ba13-d4ddb83343a0\") " pod="openstack-operators/keystone-operator-controller-manager-7f7964dd58-bjc24" Mar 11 13:38:33 crc kubenswrapper[4923]: I0311 13:38:33.621696 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7f7964dd58-bjc24" Mar 11 13:38:33 crc kubenswrapper[4923]: I0311 13:38:33.797626 4923 scope.go:117] "RemoveContainer" containerID="462bd3b780876dcdd389b70651d32cd1355063a305d760959bd250f0208c97fa" Mar 11 13:38:34 crc kubenswrapper[4923]: I0311 13:38:34.015728 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7f7964dd58-bjc24"] Mar 11 13:38:34 crc kubenswrapper[4923]: W0311 13:38:34.025508 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda492ddb6_a936_4f7d_ba13_d4ddb83343a0.slice/crio-54bfe8d9f99c53214c39a1914bff1240f2bd28ab9d71eb8ac8fa6fb1ca2b0c3f WatchSource:0}: Error finding container 54bfe8d9f99c53214c39a1914bff1240f2bd28ab9d71eb8ac8fa6fb1ca2b0c3f: Status 404 returned error can't find the container with id 54bfe8d9f99c53214c39a1914bff1240f2bd28ab9d71eb8ac8fa6fb1ca2b0c3f Mar 11 13:38:34 crc kubenswrapper[4923]: I0311 13:38:34.631197 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7f7964dd58-bjc24" event={"ID":"a492ddb6-a936-4f7d-ba13-d4ddb83343a0","Type":"ContainerStarted","Data":"54bfe8d9f99c53214c39a1914bff1240f2bd28ab9d71eb8ac8fa6fb1ca2b0c3f"} Mar 11 13:38:37 crc kubenswrapper[4923]: I0311 13:38:37.654243 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7f7964dd58-bjc24" event={"ID":"a492ddb6-a936-4f7d-ba13-d4ddb83343a0","Type":"ContainerStarted","Data":"361b8cd964aa569030f644308b42341bc88227eb11c98ffbcebb667c5bb597cc"} Mar 11 13:38:37 crc kubenswrapper[4923]: I0311 13:38:37.655886 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7f7964dd58-bjc24" Mar 11 13:38:37 crc kubenswrapper[4923]: I0311 13:38:37.675613 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7f7964dd58-bjc24" podStartSLOduration=1.3582216200000001 podStartE2EDuration="4.675590989s" podCreationTimestamp="2026-03-11 13:38:33 +0000 UTC" firstStartedPulling="2026-03-11 13:38:34.029675284 +0000 UTC m=+961.711576998" lastFinishedPulling="2026-03-11 13:38:37.347044653 +0000 UTC m=+965.028946367" observedRunningTime="2026-03-11 13:38:37.675005453 +0000 UTC m=+965.356907167" watchObservedRunningTime="2026-03-11 13:38:37.675590989 +0000 UTC m=+965.357492713" Mar 11 13:38:40 crc kubenswrapper[4923]: I0311 13:38:40.613215 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x9clt"] Mar 11 13:38:40 crc kubenswrapper[4923]: I0311 13:38:40.615021 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x9clt" Mar 11 13:38:40 crc kubenswrapper[4923]: I0311 13:38:40.626091 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x9clt"] Mar 11 13:38:40 crc kubenswrapper[4923]: I0311 13:38:40.685751 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1df5f1d-635a-498d-8310-5c8a67181bea-catalog-content\") pod \"redhat-operators-x9clt\" (UID: \"d1df5f1d-635a-498d-8310-5c8a67181bea\") " pod="openshift-marketplace/redhat-operators-x9clt" Mar 11 13:38:40 crc kubenswrapper[4923]: I0311 13:38:40.685804 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1df5f1d-635a-498d-8310-5c8a67181bea-utilities\") pod \"redhat-operators-x9clt\" (UID: \"d1df5f1d-635a-498d-8310-5c8a67181bea\") " pod="openshift-marketplace/redhat-operators-x9clt" Mar 11 13:38:40 crc kubenswrapper[4923]: I0311 13:38:40.685828 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqtts\" (UniqueName: \"kubernetes.io/projected/d1df5f1d-635a-498d-8310-5c8a67181bea-kube-api-access-dqtts\") pod \"redhat-operators-x9clt\" (UID: \"d1df5f1d-635a-498d-8310-5c8a67181bea\") " pod="openshift-marketplace/redhat-operators-x9clt" Mar 11 13:38:40 crc kubenswrapper[4923]: I0311 13:38:40.787283 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1df5f1d-635a-498d-8310-5c8a67181bea-catalog-content\") pod \"redhat-operators-x9clt\" (UID: \"d1df5f1d-635a-498d-8310-5c8a67181bea\") " pod="openshift-marketplace/redhat-operators-x9clt" Mar 11 13:38:40 crc kubenswrapper[4923]: I0311 13:38:40.787416 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1df5f1d-635a-498d-8310-5c8a67181bea-utilities\") pod \"redhat-operators-x9clt\" (UID: \"d1df5f1d-635a-498d-8310-5c8a67181bea\") " pod="openshift-marketplace/redhat-operators-x9clt" Mar 11 13:38:40 crc kubenswrapper[4923]: I0311 13:38:40.787471 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqtts\" (UniqueName: \"kubernetes.io/projected/d1df5f1d-635a-498d-8310-5c8a67181bea-kube-api-access-dqtts\") pod \"redhat-operators-x9clt\" (UID: \"d1df5f1d-635a-498d-8310-5c8a67181bea\") " pod="openshift-marketplace/redhat-operators-x9clt" Mar 11 13:38:40 crc kubenswrapper[4923]: I0311 13:38:40.787968 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1df5f1d-635a-498d-8310-5c8a67181bea-catalog-content\") pod \"redhat-operators-x9clt\" (UID: \"d1df5f1d-635a-498d-8310-5c8a67181bea\") " pod="openshift-marketplace/redhat-operators-x9clt" Mar 11 13:38:40 crc kubenswrapper[4923]: I0311 13:38:40.788004 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1df5f1d-635a-498d-8310-5c8a67181bea-utilities\") pod \"redhat-operators-x9clt\" (UID: \"d1df5f1d-635a-498d-8310-5c8a67181bea\") " pod="openshift-marketplace/redhat-operators-x9clt" Mar 11 13:38:40 crc kubenswrapper[4923]: I0311 13:38:40.811012 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqtts\" (UniqueName: \"kubernetes.io/projected/d1df5f1d-635a-498d-8310-5c8a67181bea-kube-api-access-dqtts\") pod \"redhat-operators-x9clt\" (UID: \"d1df5f1d-635a-498d-8310-5c8a67181bea\") " pod="openshift-marketplace/redhat-operators-x9clt" Mar 11 13:38:40 crc kubenswrapper[4923]: I0311 13:38:40.932991 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x9clt" Mar 11 13:38:41 crc kubenswrapper[4923]: I0311 13:38:41.408614 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x9clt"] Mar 11 13:38:41 crc kubenswrapper[4923]: W0311 13:38:41.418548 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1df5f1d_635a_498d_8310_5c8a67181bea.slice/crio-0a3beba7a13246da29a3a8c61885b7e8f704caf3975d2a479555d5bc6149a8d6 WatchSource:0}: Error finding container 0a3beba7a13246da29a3a8c61885b7e8f704caf3975d2a479555d5bc6149a8d6: Status 404 returned error can't find the container with id 0a3beba7a13246da29a3a8c61885b7e8f704caf3975d2a479555d5bc6149a8d6 Mar 11 13:38:41 crc kubenswrapper[4923]: I0311 13:38:41.683320 4923 generic.go:334] "Generic (PLEG): container finished" podID="d1df5f1d-635a-498d-8310-5c8a67181bea" containerID="455a6f5fb33174fd1ab20fa90a0830f61c86c86a7806f8c6c64531f9268f54cb" exitCode=0 Mar 11 13:38:41 crc kubenswrapper[4923]: I0311 13:38:41.683381 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9clt" event={"ID":"d1df5f1d-635a-498d-8310-5c8a67181bea","Type":"ContainerDied","Data":"455a6f5fb33174fd1ab20fa90a0830f61c86c86a7806f8c6c64531f9268f54cb"} Mar 11 13:38:41 crc kubenswrapper[4923]: I0311 13:38:41.683406 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9clt" event={"ID":"d1df5f1d-635a-498d-8310-5c8a67181bea","Type":"ContainerStarted","Data":"0a3beba7a13246da29a3a8c61885b7e8f704caf3975d2a479555d5bc6149a8d6"} Mar 11 13:38:43 crc kubenswrapper[4923]: I0311 13:38:43.628772 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7f7964dd58-bjc24" Mar 11 13:38:43 crc kubenswrapper[4923]: I0311 13:38:43.703014 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9clt" event={"ID":"d1df5f1d-635a-498d-8310-5c8a67181bea","Type":"ContainerStarted","Data":"0ffed885c97da567865dbe4bdd229647f788ce983c2d21f7af16acb7904afcd1"} Mar 11 13:38:44 crc kubenswrapper[4923]: I0311 13:38:44.714517 4923 generic.go:334] "Generic (PLEG): container finished" podID="50119944-78f3-4b19-b19b-1bc87e93e3fc" containerID="a4d278c08d19fb41723a97321fa1ccb51ebe02c480e860f8ff4dd9b8e2b1c166" exitCode=0 Mar 11 13:38:44 crc kubenswrapper[4923]: I0311 13:38:44.714603 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"50119944-78f3-4b19-b19b-1bc87e93e3fc","Type":"ContainerDied","Data":"a4d278c08d19fb41723a97321fa1ccb51ebe02c480e860f8ff4dd9b8e2b1c166"} Mar 11 13:38:44 crc kubenswrapper[4923]: I0311 13:38:44.717976 4923 generic.go:334] "Generic (PLEG): container finished" podID="d1df5f1d-635a-498d-8310-5c8a67181bea" containerID="0ffed885c97da567865dbe4bdd229647f788ce983c2d21f7af16acb7904afcd1" exitCode=0 Mar 11 13:38:44 crc kubenswrapper[4923]: I0311 13:38:44.718021 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9clt" event={"ID":"d1df5f1d-635a-498d-8310-5c8a67181bea","Type":"ContainerDied","Data":"0ffed885c97da567865dbe4bdd229647f788ce983c2d21f7af16acb7904afcd1"} Mar 11 13:38:45 crc kubenswrapper[4923]: I0311 13:38:45.725948 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/rabbitmq-server-0" event={"ID":"50119944-78f3-4b19-b19b-1bc87e93e3fc","Type":"ContainerStarted","Data":"c98e79db4dd8af7905144c89c1fe102156f454adc4b3752e69b23a3f536c6b6b"} Mar 11 13:38:45 crc kubenswrapper[4923]: I0311 13:38:45.727545 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:45 crc kubenswrapper[4923]: I0311 13:38:45.730035 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9clt" event={"ID":"d1df5f1d-635a-498d-8310-5c8a67181bea","Type":"ContainerStarted","Data":"e3a78f3446a526a9af7f72645ec05e530d27e89a7788ce83c5820470e879b446"} Mar 11 13:38:45 crc kubenswrapper[4923]: I0311 13:38:45.750405 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/rabbitmq-server-0" podStartSLOduration=37.926484046 podStartE2EDuration="43.750387728s" podCreationTimestamp="2026-03-11 13:38:02 +0000 UTC" firstStartedPulling="2026-03-11 13:38:04.64073901 +0000 UTC m=+932.322640724" lastFinishedPulling="2026-03-11 13:38:10.464642682 +0000 UTC m=+938.146544406" observedRunningTime="2026-03-11 13:38:45.7472173 +0000 UTC m=+973.429119024" watchObservedRunningTime="2026-03-11 13:38:45.750387728 +0000 UTC m=+973.432289442" Mar 11 13:38:45 crc kubenswrapper[4923]: I0311 13:38:45.777104 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x9clt" podStartSLOduration=2.258631876 podStartE2EDuration="5.777087442s" podCreationTimestamp="2026-03-11 13:38:40 +0000 UTC" firstStartedPulling="2026-03-11 13:38:41.684574001 +0000 UTC m=+969.366475715" lastFinishedPulling="2026-03-11 13:38:45.203029567 +0000 UTC m=+972.884931281" observedRunningTime="2026-03-11 13:38:45.776330322 +0000 UTC m=+973.458232036" watchObservedRunningTime="2026-03-11 13:38:45.777087442 +0000 UTC m=+973.458989156" Mar 11 13:38:47 crc kubenswrapper[4923]: I0311 13:38:47.935595 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-db-create-clm8k"] Mar 11 13:38:47 crc kubenswrapper[4923]: I0311 13:38:47.936744 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-create-clm8k" Mar 11 13:38:47 crc kubenswrapper[4923]: I0311 13:38:47.945298 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-f524-account-create-update-cdc7k"] Mar 11 13:38:47 crc kubenswrapper[4923]: I0311 13:38:47.946276 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-f524-account-create-update-cdc7k" Mar 11 13:38:47 crc kubenswrapper[4923]: I0311 13:38:47.948446 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-db-secret" Mar 11 13:38:47 crc kubenswrapper[4923]: I0311 13:38:47.950608 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-db-create-clm8k"] Mar 11 13:38:47 crc kubenswrapper[4923]: I0311 13:38:47.992411 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-f524-account-create-update-cdc7k"] Mar 11 13:38:48 crc kubenswrapper[4923]: I0311 13:38:48.113909 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19fc67d1-7c03-4a55-8baa-362e5a8a2344-operator-scripts\") pod \"keystone-f524-account-create-update-cdc7k\" (UID: \"19fc67d1-7c03-4a55-8baa-362e5a8a2344\") " pod="cinder-kuttl-tests/keystone-f524-account-create-update-cdc7k" Mar 11 13:38:48 crc kubenswrapper[4923]: I0311 13:38:48.114309 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnd4q\" (UniqueName: \"kubernetes.io/projected/19fc67d1-7c03-4a55-8baa-362e5a8a2344-kube-api-access-gnd4q\") pod \"keystone-f524-account-create-update-cdc7k\" (UID: \"19fc67d1-7c03-4a55-8baa-362e5a8a2344\") " pod="cinder-kuttl-tests/keystone-f524-account-create-update-cdc7k" Mar 11 13:38:48 crc kubenswrapper[4923]: I0311 13:38:48.114358 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60b9b9dd-cd32-48bb-9995-36bf024f9087-operator-scripts\") pod \"keystone-db-create-clm8k\" (UID: \"60b9b9dd-cd32-48bb-9995-36bf024f9087\") " pod="cinder-kuttl-tests/keystone-db-create-clm8k" Mar 11 13:38:48 crc kubenswrapper[4923]: I0311 13:38:48.114439 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5tfn\" (UniqueName: \"kubernetes.io/projected/60b9b9dd-cd32-48bb-9995-36bf024f9087-kube-api-access-c5tfn\") pod \"keystone-db-create-clm8k\" (UID: \"60b9b9dd-cd32-48bb-9995-36bf024f9087\") " pod="cinder-kuttl-tests/keystone-db-create-clm8k" Mar 11 13:38:48 crc kubenswrapper[4923]: I0311 13:38:48.215259 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19fc67d1-7c03-4a55-8baa-362e5a8a2344-operator-scripts\") pod \"keystone-f524-account-create-update-cdc7k\" (UID: \"19fc67d1-7c03-4a55-8baa-362e5a8a2344\") " pod="cinder-kuttl-tests/keystone-f524-account-create-update-cdc7k" Mar 11 13:38:48 crc kubenswrapper[4923]: I0311 13:38:48.215370 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnd4q\" (UniqueName: \"kubernetes.io/projected/19fc67d1-7c03-4a55-8baa-362e5a8a2344-kube-api-access-gnd4q\") pod \"keystone-f524-account-create-update-cdc7k\" (UID: \"19fc67d1-7c03-4a55-8baa-362e5a8a2344\") " pod="cinder-kuttl-tests/keystone-f524-account-create-update-cdc7k" Mar 11 13:38:48 crc kubenswrapper[4923]: I0311 13:38:48.215423 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60b9b9dd-cd32-48bb-9995-36bf024f9087-operator-scripts\") pod \"keystone-db-create-clm8k\" (UID: \"60b9b9dd-cd32-48bb-9995-36bf024f9087\") " pod="cinder-kuttl-tests/keystone-db-create-clm8k" Mar 11 13:38:48 crc kubenswrapper[4923]: I0311 13:38:48.215550 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5tfn\" (UniqueName: \"kubernetes.io/projected/60b9b9dd-cd32-48bb-9995-36bf024f9087-kube-api-access-c5tfn\") pod \"keystone-db-create-clm8k\" (UID: \"60b9b9dd-cd32-48bb-9995-36bf024f9087\") " pod="cinder-kuttl-tests/keystone-db-create-clm8k" Mar 11 13:38:48 crc kubenswrapper[4923]: I0311 13:38:48.216847 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19fc67d1-7c03-4a55-8baa-362e5a8a2344-operator-scripts\") pod \"keystone-f524-account-create-update-cdc7k\" (UID: \"19fc67d1-7c03-4a55-8baa-362e5a8a2344\") " pod="cinder-kuttl-tests/keystone-f524-account-create-update-cdc7k" Mar 11 13:38:48 crc kubenswrapper[4923]: I0311 13:38:48.217769 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60b9b9dd-cd32-48bb-9995-36bf024f9087-operator-scripts\") pod \"keystone-db-create-clm8k\" (UID: \"60b9b9dd-cd32-48bb-9995-36bf024f9087\") " pod="cinder-kuttl-tests/keystone-db-create-clm8k" Mar 11 13:38:48 crc kubenswrapper[4923]: I0311 13:38:48.239146 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5tfn\" (UniqueName: \"kubernetes.io/projected/60b9b9dd-cd32-48bb-9995-36bf024f9087-kube-api-access-c5tfn\") pod \"keystone-db-create-clm8k\" (UID: \"60b9b9dd-cd32-48bb-9995-36bf024f9087\") " pod="cinder-kuttl-tests/keystone-db-create-clm8k" Mar 11 13:38:48 crc kubenswrapper[4923]: I0311 13:38:48.239680 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnd4q\" (UniqueName: \"kubernetes.io/projected/19fc67d1-7c03-4a55-8baa-362e5a8a2344-kube-api-access-gnd4q\") pod \"keystone-f524-account-create-update-cdc7k\" (UID: \"19fc67d1-7c03-4a55-8baa-362e5a8a2344\") " pod="cinder-kuttl-tests/keystone-f524-account-create-update-cdc7k" Mar 11 13:38:48 crc kubenswrapper[4923]: I0311 13:38:48.251891 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-create-clm8k" Mar 11 13:38:48 crc kubenswrapper[4923]: I0311 13:38:48.280745 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-f524-account-create-update-cdc7k" Mar 11 13:38:48 crc kubenswrapper[4923]: I0311 13:38:48.739451 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-db-create-clm8k"] Mar 11 13:38:48 crc kubenswrapper[4923]: W0311 13:38:48.811619 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod19fc67d1_7c03_4a55_8baa_362e5a8a2344.slice/crio-8d8fe430e70157a41d00981bb90b71d7d0f8817a1351649adb11945fd6c374cc WatchSource:0}: Error finding container 8d8fe430e70157a41d00981bb90b71d7d0f8817a1351649adb11945fd6c374cc: Status 404 returned error can't find the container with id 8d8fe430e70157a41d00981bb90b71d7d0f8817a1351649adb11945fd6c374cc Mar 11 13:38:48 crc kubenswrapper[4923]: I0311 13:38:48.813119 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-f524-account-create-update-cdc7k"] Mar 11 13:38:49 crc kubenswrapper[4923]: I0311 13:38:49.756009 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-create-clm8k" event={"ID":"60b9b9dd-cd32-48bb-9995-36bf024f9087","Type":"ContainerStarted","Data":"a7d1a0e725782476084f70d25781444849a26204d786d6bc8c3e9cea9397ce6c"} Mar 11 13:38:49 crc kubenswrapper[4923]: I0311 13:38:49.756899 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-create-clm8k" event={"ID":"60b9b9dd-cd32-48bb-9995-36bf024f9087","Type":"ContainerStarted","Data":"84fd76d017cfe634af19937ad98b6b62773967200d9c2f2ab7371ad94474a2ee"} Mar 11 13:38:49 crc kubenswrapper[4923]: I0311 13:38:49.758822 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-f524-account-create-update-cdc7k" event={"ID":"19fc67d1-7c03-4a55-8baa-362e5a8a2344","Type":"ContainerStarted","Data":"0848cd26093c6a84dbeea86a0c14b063883ca2130c21912db2187eba4e273c4b"} Mar 11 13:38:49 crc kubenswrapper[4923]: I0311 13:38:49.758879 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-f524-account-create-update-cdc7k" event={"ID":"19fc67d1-7c03-4a55-8baa-362e5a8a2344","Type":"ContainerStarted","Data":"8d8fe430e70157a41d00981bb90b71d7d0f8817a1351649adb11945fd6c374cc"} Mar 11 13:38:49 crc kubenswrapper[4923]: I0311 13:38:49.793544 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/keystone-f524-account-create-update-cdc7k" podStartSLOduration=2.793517976 podStartE2EDuration="2.793517976s" podCreationTimestamp="2026-03-11 13:38:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:38:49.788046537 +0000 UTC m=+977.469948251" watchObservedRunningTime="2026-03-11 13:38:49.793517976 +0000 UTC m=+977.475419710" Mar 11 13:38:50 crc kubenswrapper[4923]: I0311 13:38:50.765921 4923 generic.go:334] "Generic (PLEG): container finished" podID="19fc67d1-7c03-4a55-8baa-362e5a8a2344" containerID="0848cd26093c6a84dbeea86a0c14b063883ca2130c21912db2187eba4e273c4b" exitCode=0 Mar 11 13:38:50 crc kubenswrapper[4923]: I0311 13:38:50.765997 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-f524-account-create-update-cdc7k" event={"ID":"19fc67d1-7c03-4a55-8baa-362e5a8a2344","Type":"ContainerDied","Data":"0848cd26093c6a84dbeea86a0c14b063883ca2130c21912db2187eba4e273c4b"} Mar 11 13:38:50 crc kubenswrapper[4923]: I0311 13:38:50.767698 4923 generic.go:334] "Generic (PLEG): container finished" podID="60b9b9dd-cd32-48bb-9995-36bf024f9087" containerID="a7d1a0e725782476084f70d25781444849a26204d786d6bc8c3e9cea9397ce6c" exitCode=0 Mar 11 13:38:50 crc kubenswrapper[4923]: I0311 13:38:50.767727 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-create-clm8k" event={"ID":"60b9b9dd-cd32-48bb-9995-36bf024f9087","Type":"ContainerDied","Data":"a7d1a0e725782476084f70d25781444849a26204d786d6bc8c3e9cea9397ce6c"} Mar 11 13:38:50 crc kubenswrapper[4923]: I0311 13:38:50.934223 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x9clt" Mar 11 13:38:50 crc kubenswrapper[4923]: I0311 13:38:50.934606 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x9clt" Mar 11 13:38:51 crc kubenswrapper[4923]: I0311 13:38:51.133287 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-create-clm8k" Mar 11 13:38:51 crc kubenswrapper[4923]: I0311 13:38:51.292909 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5tfn\" (UniqueName: \"kubernetes.io/projected/60b9b9dd-cd32-48bb-9995-36bf024f9087-kube-api-access-c5tfn\") pod \"60b9b9dd-cd32-48bb-9995-36bf024f9087\" (UID: \"60b9b9dd-cd32-48bb-9995-36bf024f9087\") " Mar 11 13:38:51 crc kubenswrapper[4923]: I0311 13:38:51.293029 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60b9b9dd-cd32-48bb-9995-36bf024f9087-operator-scripts\") pod \"60b9b9dd-cd32-48bb-9995-36bf024f9087\" (UID: \"60b9b9dd-cd32-48bb-9995-36bf024f9087\") " Mar 11 13:38:51 crc kubenswrapper[4923]: I0311 13:38:51.293663 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60b9b9dd-cd32-48bb-9995-36bf024f9087-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "60b9b9dd-cd32-48bb-9995-36bf024f9087" (UID: "60b9b9dd-cd32-48bb-9995-36bf024f9087"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:38:51 crc kubenswrapper[4923]: I0311 13:38:51.299603 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60b9b9dd-cd32-48bb-9995-36bf024f9087-kube-api-access-c5tfn" (OuterVolumeSpecName: "kube-api-access-c5tfn") pod "60b9b9dd-cd32-48bb-9995-36bf024f9087" (UID: "60b9b9dd-cd32-48bb-9995-36bf024f9087"). InnerVolumeSpecName "kube-api-access-c5tfn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:38:51 crc kubenswrapper[4923]: I0311 13:38:51.394592 4923 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/60b9b9dd-cd32-48bb-9995-36bf024f9087-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 11 13:38:51 crc kubenswrapper[4923]: I0311 13:38:51.394635 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5tfn\" (UniqueName: \"kubernetes.io/projected/60b9b9dd-cd32-48bb-9995-36bf024f9087-kube-api-access-c5tfn\") on node \"crc\" DevicePath \"\"" Mar 11 13:38:51 crc kubenswrapper[4923]: I0311 13:38:51.776639 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-create-clm8k" Mar 11 13:38:51 crc kubenswrapper[4923]: I0311 13:38:51.779553 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-create-clm8k" event={"ID":"60b9b9dd-cd32-48bb-9995-36bf024f9087","Type":"ContainerDied","Data":"84fd76d017cfe634af19937ad98b6b62773967200d9c2f2ab7371ad94474a2ee"} Mar 11 13:38:51 crc kubenswrapper[4923]: I0311 13:38:51.779578 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84fd76d017cfe634af19937ad98b6b62773967200d9c2f2ab7371ad94474a2ee" Mar 11 13:38:51 crc kubenswrapper[4923]: I0311 13:38:51.797461 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-index-66xkx"] Mar 11 13:38:51 crc kubenswrapper[4923]: E0311 13:38:51.797723 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60b9b9dd-cd32-48bb-9995-36bf024f9087" containerName="mariadb-database-create" Mar 11 13:38:51 crc kubenswrapper[4923]: I0311 13:38:51.797743 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="60b9b9dd-cd32-48bb-9995-36bf024f9087" containerName="mariadb-database-create" Mar 11 13:38:51 crc kubenswrapper[4923]: I0311 13:38:51.797905 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="60b9b9dd-cd32-48bb-9995-36bf024f9087" containerName="mariadb-database-create" Mar 11 13:38:51 crc kubenswrapper[4923]: I0311 13:38:51.798416 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-index-66xkx" Mar 11 13:38:51 crc kubenswrapper[4923]: I0311 13:38:51.801716 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-index-dockercfg-sbp59" Mar 11 13:38:51 crc kubenswrapper[4923]: I0311 13:38:51.809823 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-index-66xkx"] Mar 11 13:38:51 crc kubenswrapper[4923]: I0311 13:38:51.900473 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cs4gg\" (UniqueName: \"kubernetes.io/projected/4c504c2c-5548-406f-b2f9-55f53c126a0f-kube-api-access-cs4gg\") pod \"cinder-operator-index-66xkx\" (UID: \"4c504c2c-5548-406f-b2f9-55f53c126a0f\") " pod="openstack-operators/cinder-operator-index-66xkx" Mar 11 13:38:51 crc kubenswrapper[4923]: I0311 13:38:51.994508 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x9clt" podUID="d1df5f1d-635a-498d-8310-5c8a67181bea" containerName="registry-server" probeResult="failure" output=< Mar 11 13:38:51 crc kubenswrapper[4923]: timeout: failed to connect service ":50051" within 1s Mar 11 13:38:51 crc kubenswrapper[4923]: > Mar 11 13:38:52 crc kubenswrapper[4923]: I0311 13:38:52.001644 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cs4gg\" (UniqueName: \"kubernetes.io/projected/4c504c2c-5548-406f-b2f9-55f53c126a0f-kube-api-access-cs4gg\") pod \"cinder-operator-index-66xkx\" (UID: \"4c504c2c-5548-406f-b2f9-55f53c126a0f\") " pod="openstack-operators/cinder-operator-index-66xkx" Mar 11 13:38:52 crc kubenswrapper[4923]: I0311 13:38:52.032761 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cs4gg\" (UniqueName: \"kubernetes.io/projected/4c504c2c-5548-406f-b2f9-55f53c126a0f-kube-api-access-cs4gg\") pod \"cinder-operator-index-66xkx\" (UID: \"4c504c2c-5548-406f-b2f9-55f53c126a0f\") " pod="openstack-operators/cinder-operator-index-66xkx" Mar 11 13:38:52 crc kubenswrapper[4923]: I0311 13:38:52.116153 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-f524-account-create-update-cdc7k" Mar 11 13:38:52 crc kubenswrapper[4923]: I0311 13:38:52.126107 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-index-66xkx" Mar 11 13:38:52 crc kubenswrapper[4923]: I0311 13:38:52.304811 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19fc67d1-7c03-4a55-8baa-362e5a8a2344-operator-scripts\") pod \"19fc67d1-7c03-4a55-8baa-362e5a8a2344\" (UID: \"19fc67d1-7c03-4a55-8baa-362e5a8a2344\") " Mar 11 13:38:52 crc kubenswrapper[4923]: I0311 13:38:52.305224 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnd4q\" (UniqueName: \"kubernetes.io/projected/19fc67d1-7c03-4a55-8baa-362e5a8a2344-kube-api-access-gnd4q\") pod \"19fc67d1-7c03-4a55-8baa-362e5a8a2344\" (UID: \"19fc67d1-7c03-4a55-8baa-362e5a8a2344\") " Mar 11 13:38:52 crc kubenswrapper[4923]: I0311 13:38:52.305621 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19fc67d1-7c03-4a55-8baa-362e5a8a2344-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "19fc67d1-7c03-4a55-8baa-362e5a8a2344" (UID: "19fc67d1-7c03-4a55-8baa-362e5a8a2344"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:38:52 crc kubenswrapper[4923]: I0311 13:38:52.319547 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19fc67d1-7c03-4a55-8baa-362e5a8a2344-kube-api-access-gnd4q" (OuterVolumeSpecName: "kube-api-access-gnd4q") pod "19fc67d1-7c03-4a55-8baa-362e5a8a2344" (UID: "19fc67d1-7c03-4a55-8baa-362e5a8a2344"). InnerVolumeSpecName "kube-api-access-gnd4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:38:52 crc kubenswrapper[4923]: I0311 13:38:52.406550 4923 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/19fc67d1-7c03-4a55-8baa-362e5a8a2344-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 11 13:38:52 crc kubenswrapper[4923]: I0311 13:38:52.406577 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnd4q\" (UniqueName: \"kubernetes.io/projected/19fc67d1-7c03-4a55-8baa-362e5a8a2344-kube-api-access-gnd4q\") on node \"crc\" DevicePath \"\"" Mar 11 13:38:52 crc kubenswrapper[4923]: I0311 13:38:52.614905 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-index-66xkx"] Mar 11 13:38:52 crc kubenswrapper[4923]: I0311 13:38:52.788514 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-index-66xkx" event={"ID":"4c504c2c-5548-406f-b2f9-55f53c126a0f","Type":"ContainerStarted","Data":"2b24f92b29c5f7016a12d1c997b1ba62ba11bc38699f8844d386c39ba96b1a91"} Mar 11 13:38:52 crc kubenswrapper[4923]: I0311 13:38:52.791431 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-f524-account-create-update-cdc7k" event={"ID":"19fc67d1-7c03-4a55-8baa-362e5a8a2344","Type":"ContainerDied","Data":"8d8fe430e70157a41d00981bb90b71d7d0f8817a1351649adb11945fd6c374cc"} Mar 11 13:38:52 crc kubenswrapper[4923]: I0311 13:38:52.791481 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d8fe430e70157a41d00981bb90b71d7d0f8817a1351649adb11945fd6c374cc" Mar 11 13:38:52 crc kubenswrapper[4923]: I0311 13:38:52.791603 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-f524-account-create-update-cdc7k" Mar 11 13:38:54 crc kubenswrapper[4923]: I0311 13:38:54.091958 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/rabbitmq-server-0" Mar 11 13:38:55 crc kubenswrapper[4923]: I0311 13:38:55.830543 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-index-66xkx" event={"ID":"4c504c2c-5548-406f-b2f9-55f53c126a0f","Type":"ContainerStarted","Data":"9bc57002e0a8c13e2526b4b4b9750e2057331f0940430c13c98db7927fa13695"} Mar 11 13:38:56 crc kubenswrapper[4923]: I0311 13:38:56.858774 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-index-66xkx" podStartSLOduration=3.429816191 podStartE2EDuration="5.858747041s" podCreationTimestamp="2026-03-11 13:38:51 +0000 UTC" firstStartedPulling="2026-03-11 13:38:52.627087631 +0000 UTC m=+980.308989385" lastFinishedPulling="2026-03-11 13:38:55.056018521 +0000 UTC m=+982.737920235" observedRunningTime="2026-03-11 13:38:56.849443447 +0000 UTC m=+984.531345201" watchObservedRunningTime="2026-03-11 13:38:56.858747041 +0000 UTC m=+984.540648805" Mar 11 13:38:58 crc kubenswrapper[4923]: I0311 13:38:58.540122 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-cjv6x"] Mar 11 13:38:58 crc kubenswrapper[4923]: E0311 13:38:58.540622 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19fc67d1-7c03-4a55-8baa-362e5a8a2344" containerName="mariadb-account-create-update" Mar 11 13:38:58 crc kubenswrapper[4923]: I0311 13:38:58.540656 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="19fc67d1-7c03-4a55-8baa-362e5a8a2344" containerName="mariadb-account-create-update" Mar 11 13:38:58 crc kubenswrapper[4923]: I0311 13:38:58.540894 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="19fc67d1-7c03-4a55-8baa-362e5a8a2344" containerName="mariadb-account-create-update" Mar 11 13:38:58 crc kubenswrapper[4923]: I0311 13:38:58.541671 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-sync-cjv6x" Mar 11 13:38:58 crc kubenswrapper[4923]: I0311 13:38:58.544623 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone" Mar 11 13:38:58 crc kubenswrapper[4923]: I0311 13:38:58.545590 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-keystone-dockercfg-7wv7t" Mar 11 13:38:58 crc kubenswrapper[4923]: I0311 13:38:58.546447 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-config-data" Mar 11 13:38:58 crc kubenswrapper[4923]: I0311 13:38:58.546763 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-scripts" Mar 11 13:38:58 crc kubenswrapper[4923]: I0311 13:38:58.552796 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-cjv6x"] Mar 11 13:38:58 crc kubenswrapper[4923]: I0311 13:38:58.699832 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fq5zx\" (UniqueName: \"kubernetes.io/projected/eae8474f-7523-4ee9-8f02-7be328688043-kube-api-access-fq5zx\") pod \"keystone-db-sync-cjv6x\" (UID: \"eae8474f-7523-4ee9-8f02-7be328688043\") " pod="cinder-kuttl-tests/keystone-db-sync-cjv6x" Mar 11 13:38:58 crc kubenswrapper[4923]: I0311 13:38:58.699989 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eae8474f-7523-4ee9-8f02-7be328688043-config-data\") pod \"keystone-db-sync-cjv6x\" (UID: \"eae8474f-7523-4ee9-8f02-7be328688043\") " pod="cinder-kuttl-tests/keystone-db-sync-cjv6x" Mar 11 13:38:58 crc kubenswrapper[4923]: I0311 13:38:58.801184 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eae8474f-7523-4ee9-8f02-7be328688043-config-data\") pod \"keystone-db-sync-cjv6x\" (UID: \"eae8474f-7523-4ee9-8f02-7be328688043\") " pod="cinder-kuttl-tests/keystone-db-sync-cjv6x" Mar 11 13:38:58 crc kubenswrapper[4923]: I0311 13:38:58.801294 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fq5zx\" (UniqueName: \"kubernetes.io/projected/eae8474f-7523-4ee9-8f02-7be328688043-kube-api-access-fq5zx\") pod \"keystone-db-sync-cjv6x\" (UID: \"eae8474f-7523-4ee9-8f02-7be328688043\") " pod="cinder-kuttl-tests/keystone-db-sync-cjv6x" Mar 11 13:38:58 crc kubenswrapper[4923]: I0311 13:38:58.812906 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eae8474f-7523-4ee9-8f02-7be328688043-config-data\") pod \"keystone-db-sync-cjv6x\" (UID: \"eae8474f-7523-4ee9-8f02-7be328688043\") " pod="cinder-kuttl-tests/keystone-db-sync-cjv6x" Mar 11 13:38:58 crc kubenswrapper[4923]: I0311 13:38:58.833417 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fq5zx\" (UniqueName: \"kubernetes.io/projected/eae8474f-7523-4ee9-8f02-7be328688043-kube-api-access-fq5zx\") pod \"keystone-db-sync-cjv6x\" (UID: \"eae8474f-7523-4ee9-8f02-7be328688043\") " pod="cinder-kuttl-tests/keystone-db-sync-cjv6x" Mar 11 13:38:58 crc kubenswrapper[4923]: I0311 13:38:58.882501 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-sync-cjv6x" Mar 11 13:38:59 crc kubenswrapper[4923]: I0311 13:38:59.353572 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-cjv6x"] Mar 11 13:38:59 crc kubenswrapper[4923]: W0311 13:38:59.356277 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeae8474f_7523_4ee9_8f02_7be328688043.slice/crio-6dda9fe864106188b3349b257fb6022c9c2f5495fea427307b1f469904617cf7 WatchSource:0}: Error finding container 6dda9fe864106188b3349b257fb6022c9c2f5495fea427307b1f469904617cf7: Status 404 returned error can't find the container with id 6dda9fe864106188b3349b257fb6022c9c2f5495fea427307b1f469904617cf7 Mar 11 13:38:59 crc kubenswrapper[4923]: I0311 13:38:59.862204 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-sync-cjv6x" event={"ID":"eae8474f-7523-4ee9-8f02-7be328688043","Type":"ContainerStarted","Data":"6dda9fe864106188b3349b257fb6022c9c2f5495fea427307b1f469904617cf7"} Mar 11 13:39:00 crc kubenswrapper[4923]: I0311 13:39:00.983230 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x9clt" Mar 11 13:39:01 crc kubenswrapper[4923]: I0311 13:39:01.030164 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x9clt" Mar 11 13:39:02 crc kubenswrapper[4923]: I0311 13:39:02.126799 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-index-66xkx" Mar 11 13:39:02 crc kubenswrapper[4923]: I0311 13:39:02.126859 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/cinder-operator-index-66xkx" Mar 11 13:39:02 crc kubenswrapper[4923]: I0311 13:39:02.187414 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/cinder-operator-index-66xkx" Mar 11 13:39:03 crc kubenswrapper[4923]: I0311 13:39:03.137723 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-index-66xkx" Mar 11 13:39:05 crc kubenswrapper[4923]: I0311 13:39:05.456636 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf"] Mar 11 13:39:05 crc kubenswrapper[4923]: I0311 13:39:05.459935 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf" Mar 11 13:39:05 crc kubenswrapper[4923]: I0311 13:39:05.462165 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-qkqtx" Mar 11 13:39:05 crc kubenswrapper[4923]: I0311 13:39:05.462222 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf"] Mar 11 13:39:05 crc kubenswrapper[4923]: I0311 13:39:05.545708 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5ec59e42-a1e4-4190-95bc-8370d92be5ff-util\") pod \"ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf\" (UID: \"5ec59e42-a1e4-4190-95bc-8370d92be5ff\") " pod="openstack-operators/ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf" Mar 11 13:39:05 crc kubenswrapper[4923]: I0311 13:39:05.545970 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mn2qd\" (UniqueName: \"kubernetes.io/projected/5ec59e42-a1e4-4190-95bc-8370d92be5ff-kube-api-access-mn2qd\") pod \"ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf\" (UID: \"5ec59e42-a1e4-4190-95bc-8370d92be5ff\") " pod="openstack-operators/ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf" Mar 11 13:39:05 crc kubenswrapper[4923]: I0311 13:39:05.546101 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5ec59e42-a1e4-4190-95bc-8370d92be5ff-bundle\") pod \"ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf\" (UID: \"5ec59e42-a1e4-4190-95bc-8370d92be5ff\") " pod="openstack-operators/ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf" Mar 11 13:39:05 crc kubenswrapper[4923]: I0311 13:39:05.595400 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x9clt"] Mar 11 13:39:05 crc kubenswrapper[4923]: I0311 13:39:05.596125 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x9clt" podUID="d1df5f1d-635a-498d-8310-5c8a67181bea" containerName="registry-server" containerID="cri-o://e3a78f3446a526a9af7f72645ec05e530d27e89a7788ce83c5820470e879b446" gracePeriod=2 Mar 11 13:39:05 crc kubenswrapper[4923]: I0311 13:39:05.647577 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5ec59e42-a1e4-4190-95bc-8370d92be5ff-bundle\") pod \"ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf\" (UID: \"5ec59e42-a1e4-4190-95bc-8370d92be5ff\") " pod="openstack-operators/ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf" Mar 11 13:39:05 crc kubenswrapper[4923]: I0311 13:39:05.647782 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5ec59e42-a1e4-4190-95bc-8370d92be5ff-util\") pod \"ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf\" (UID: \"5ec59e42-a1e4-4190-95bc-8370d92be5ff\") " pod="openstack-operators/ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf" Mar 11 13:39:05 crc kubenswrapper[4923]: I0311 13:39:05.647826 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mn2qd\" (UniqueName: \"kubernetes.io/projected/5ec59e42-a1e4-4190-95bc-8370d92be5ff-kube-api-access-mn2qd\") pod \"ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf\" (UID: \"5ec59e42-a1e4-4190-95bc-8370d92be5ff\") " pod="openstack-operators/ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf" Mar 11 13:39:05 crc kubenswrapper[4923]: I0311 13:39:05.648075 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5ec59e42-a1e4-4190-95bc-8370d92be5ff-bundle\") pod \"ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf\" (UID: \"5ec59e42-a1e4-4190-95bc-8370d92be5ff\") " pod="openstack-operators/ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf" Mar 11 13:39:05 crc kubenswrapper[4923]: I0311 13:39:05.648310 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5ec59e42-a1e4-4190-95bc-8370d92be5ff-util\") pod \"ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf\" (UID: \"5ec59e42-a1e4-4190-95bc-8370d92be5ff\") " pod="openstack-operators/ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf" Mar 11 13:39:05 crc kubenswrapper[4923]: I0311 13:39:05.678539 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mn2qd\" (UniqueName: \"kubernetes.io/projected/5ec59e42-a1e4-4190-95bc-8370d92be5ff-kube-api-access-mn2qd\") pod \"ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf\" (UID: \"5ec59e42-a1e4-4190-95bc-8370d92be5ff\") " pod="openstack-operators/ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf" Mar 11 13:39:05 crc kubenswrapper[4923]: I0311 13:39:05.801771 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf" Mar 11 13:39:06 crc kubenswrapper[4923]: I0311 13:39:06.131100 4923 generic.go:334] "Generic (PLEG): container finished" podID="d1df5f1d-635a-498d-8310-5c8a67181bea" containerID="e3a78f3446a526a9af7f72645ec05e530d27e89a7788ce83c5820470e879b446" exitCode=0 Mar 11 13:39:06 crc kubenswrapper[4923]: I0311 13:39:06.131151 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9clt" event={"ID":"d1df5f1d-635a-498d-8310-5c8a67181bea","Type":"ContainerDied","Data":"e3a78f3446a526a9af7f72645ec05e530d27e89a7788ce83c5820470e879b446"} Mar 11 13:39:07 crc kubenswrapper[4923]: I0311 13:39:07.293836 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x9clt" Mar 11 13:39:07 crc kubenswrapper[4923]: I0311 13:39:07.332383 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf"] Mar 11 13:39:07 crc kubenswrapper[4923]: W0311 13:39:07.339538 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5ec59e42_a1e4_4190_95bc_8370d92be5ff.slice/crio-61a271f38be773e1ec078a77fb52279b7c50836e34938ee471cf619027261b7e WatchSource:0}: Error finding container 61a271f38be773e1ec078a77fb52279b7c50836e34938ee471cf619027261b7e: Status 404 returned error can't find the container with id 61a271f38be773e1ec078a77fb52279b7c50836e34938ee471cf619027261b7e Mar 11 13:39:07 crc kubenswrapper[4923]: I0311 13:39:07.396122 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqtts\" (UniqueName: \"kubernetes.io/projected/d1df5f1d-635a-498d-8310-5c8a67181bea-kube-api-access-dqtts\") pod \"d1df5f1d-635a-498d-8310-5c8a67181bea\" (UID: \"d1df5f1d-635a-498d-8310-5c8a67181bea\") " Mar 11 13:39:07 crc kubenswrapper[4923]: I0311 13:39:07.396259 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1df5f1d-635a-498d-8310-5c8a67181bea-catalog-content\") pod \"d1df5f1d-635a-498d-8310-5c8a67181bea\" (UID: \"d1df5f1d-635a-498d-8310-5c8a67181bea\") " Mar 11 13:39:07 crc kubenswrapper[4923]: I0311 13:39:07.396323 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1df5f1d-635a-498d-8310-5c8a67181bea-utilities\") pod \"d1df5f1d-635a-498d-8310-5c8a67181bea\" (UID: \"d1df5f1d-635a-498d-8310-5c8a67181bea\") " Mar 11 13:39:07 crc kubenswrapper[4923]: I0311 13:39:07.397188 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1df5f1d-635a-498d-8310-5c8a67181bea-utilities" (OuterVolumeSpecName: "utilities") pod "d1df5f1d-635a-498d-8310-5c8a67181bea" (UID: "d1df5f1d-635a-498d-8310-5c8a67181bea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:39:07 crc kubenswrapper[4923]: I0311 13:39:07.400455 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1df5f1d-635a-498d-8310-5c8a67181bea-kube-api-access-dqtts" (OuterVolumeSpecName: "kube-api-access-dqtts") pod "d1df5f1d-635a-498d-8310-5c8a67181bea" (UID: "d1df5f1d-635a-498d-8310-5c8a67181bea"). InnerVolumeSpecName "kube-api-access-dqtts". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:39:07 crc kubenswrapper[4923]: I0311 13:39:07.497639 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1df5f1d-635a-498d-8310-5c8a67181bea-utilities\") on node \"crc\" DevicePath \"\"" Mar 11 13:39:07 crc kubenswrapper[4923]: I0311 13:39:07.497672 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqtts\" (UniqueName: \"kubernetes.io/projected/d1df5f1d-635a-498d-8310-5c8a67181bea-kube-api-access-dqtts\") on node \"crc\" DevicePath \"\"" Mar 11 13:39:07 crc kubenswrapper[4923]: I0311 13:39:07.543446 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1df5f1d-635a-498d-8310-5c8a67181bea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1df5f1d-635a-498d-8310-5c8a67181bea" (UID: "d1df5f1d-635a-498d-8310-5c8a67181bea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:39:07 crc kubenswrapper[4923]: I0311 13:39:07.598608 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1df5f1d-635a-498d-8310-5c8a67181bea-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 11 13:39:08 crc kubenswrapper[4923]: I0311 13:39:08.158160 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-sync-cjv6x" event={"ID":"eae8474f-7523-4ee9-8f02-7be328688043","Type":"ContainerStarted","Data":"45b6ed22ba4610b78ac8d31fa24b882d84c2382b59d7e2e139798e7b1ec02dfa"} Mar 11 13:39:08 crc kubenswrapper[4923]: I0311 13:39:08.159700 4923 generic.go:334] "Generic (PLEG): container finished" podID="5ec59e42-a1e4-4190-95bc-8370d92be5ff" containerID="ce5d62bdb98090e521ae6718ee3a386b9aeb164683364468f618ace92a87c5a2" exitCode=0 Mar 11 13:39:08 crc kubenswrapper[4923]: I0311 13:39:08.159797 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf" event={"ID":"5ec59e42-a1e4-4190-95bc-8370d92be5ff","Type":"ContainerDied","Data":"ce5d62bdb98090e521ae6718ee3a386b9aeb164683364468f618ace92a87c5a2"} Mar 11 13:39:08 crc kubenswrapper[4923]: I0311 13:39:08.159841 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf" event={"ID":"5ec59e42-a1e4-4190-95bc-8370d92be5ff","Type":"ContainerStarted","Data":"61a271f38be773e1ec078a77fb52279b7c50836e34938ee471cf619027261b7e"} Mar 11 13:39:08 crc kubenswrapper[4923]: I0311 13:39:08.162393 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x9clt" event={"ID":"d1df5f1d-635a-498d-8310-5c8a67181bea","Type":"ContainerDied","Data":"0a3beba7a13246da29a3a8c61885b7e8f704caf3975d2a479555d5bc6149a8d6"} Mar 11 13:39:08 crc kubenswrapper[4923]: I0311 13:39:08.162461 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x9clt" Mar 11 13:39:08 crc kubenswrapper[4923]: I0311 13:39:08.162631 4923 scope.go:117] "RemoveContainer" containerID="e3a78f3446a526a9af7f72645ec05e530d27e89a7788ce83c5820470e879b446" Mar 11 13:39:08 crc kubenswrapper[4923]: I0311 13:39:08.186966 4923 scope.go:117] "RemoveContainer" containerID="0ffed885c97da567865dbe4bdd229647f788ce983c2d21f7af16acb7904afcd1" Mar 11 13:39:08 crc kubenswrapper[4923]: I0311 13:39:08.217974 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/keystone-db-sync-cjv6x" podStartSLOduration=2.481610055 podStartE2EDuration="10.217954114s" podCreationTimestamp="2026-03-11 13:38:58 +0000 UTC" firstStartedPulling="2026-03-11 13:38:59.35955075 +0000 UTC m=+987.041452514" lastFinishedPulling="2026-03-11 13:39:07.095894859 +0000 UTC m=+994.777796573" observedRunningTime="2026-03-11 13:39:08.183563675 +0000 UTC m=+995.865465419" watchObservedRunningTime="2026-03-11 13:39:08.217954114 +0000 UTC m=+995.899855828" Mar 11 13:39:08 crc kubenswrapper[4923]: I0311 13:39:08.218170 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x9clt"] Mar 11 13:39:08 crc kubenswrapper[4923]: I0311 13:39:08.218779 4923 scope.go:117] "RemoveContainer" containerID="455a6f5fb33174fd1ab20fa90a0830f61c86c86a7806f8c6c64531f9268f54cb" Mar 11 13:39:08 crc kubenswrapper[4923]: I0311 13:39:08.235188 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x9clt"] Mar 11 13:39:09 crc kubenswrapper[4923]: I0311 13:39:09.032869 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1df5f1d-635a-498d-8310-5c8a67181bea" path="/var/lib/kubelet/pods/d1df5f1d-635a-498d-8310-5c8a67181bea/volumes" Mar 11 13:39:09 crc kubenswrapper[4923]: I0311 13:39:09.173975 4923 generic.go:334] "Generic (PLEG): container finished" podID="5ec59e42-a1e4-4190-95bc-8370d92be5ff" containerID="fb5564228534c545b70165215b5e4fc7d62695a8ed7a7a0d7c5b2034134445cc" exitCode=0 Mar 11 13:39:09 crc kubenswrapper[4923]: I0311 13:39:09.174202 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf" event={"ID":"5ec59e42-a1e4-4190-95bc-8370d92be5ff","Type":"ContainerDied","Data":"fb5564228534c545b70165215b5e4fc7d62695a8ed7a7a0d7c5b2034134445cc"} Mar 11 13:39:10 crc kubenswrapper[4923]: I0311 13:39:10.190697 4923 generic.go:334] "Generic (PLEG): container finished" podID="5ec59e42-a1e4-4190-95bc-8370d92be5ff" containerID="12fc4258b606114691c88a1b13fc43b8d0926f6d50e2707fb38db24b7a8d32ad" exitCode=0 Mar 11 13:39:10 crc kubenswrapper[4923]: I0311 13:39:10.190919 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf" event={"ID":"5ec59e42-a1e4-4190-95bc-8370d92be5ff","Type":"ContainerDied","Data":"12fc4258b606114691c88a1b13fc43b8d0926f6d50e2707fb38db24b7a8d32ad"} Mar 11 13:39:11 crc kubenswrapper[4923]: I0311 13:39:11.201052 4923 generic.go:334] "Generic (PLEG): container finished" podID="eae8474f-7523-4ee9-8f02-7be328688043" containerID="45b6ed22ba4610b78ac8d31fa24b882d84c2382b59d7e2e139798e7b1ec02dfa" exitCode=0 Mar 11 13:39:11 crc kubenswrapper[4923]: I0311 13:39:11.201145 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-sync-cjv6x" event={"ID":"eae8474f-7523-4ee9-8f02-7be328688043","Type":"ContainerDied","Data":"45b6ed22ba4610b78ac8d31fa24b882d84c2382b59d7e2e139798e7b1ec02dfa"} Mar 11 13:39:11 crc kubenswrapper[4923]: I0311 13:39:11.499905 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf" Mar 11 13:39:11 crc kubenswrapper[4923]: I0311 13:39:11.554272 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5ec59e42-a1e4-4190-95bc-8370d92be5ff-bundle\") pod \"5ec59e42-a1e4-4190-95bc-8370d92be5ff\" (UID: \"5ec59e42-a1e4-4190-95bc-8370d92be5ff\") " Mar 11 13:39:11 crc kubenswrapper[4923]: I0311 13:39:11.554417 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mn2qd\" (UniqueName: \"kubernetes.io/projected/5ec59e42-a1e4-4190-95bc-8370d92be5ff-kube-api-access-mn2qd\") pod \"5ec59e42-a1e4-4190-95bc-8370d92be5ff\" (UID: \"5ec59e42-a1e4-4190-95bc-8370d92be5ff\") " Mar 11 13:39:11 crc kubenswrapper[4923]: I0311 13:39:11.554448 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5ec59e42-a1e4-4190-95bc-8370d92be5ff-util\") pod \"5ec59e42-a1e4-4190-95bc-8370d92be5ff\" (UID: \"5ec59e42-a1e4-4190-95bc-8370d92be5ff\") " Mar 11 13:39:11 crc kubenswrapper[4923]: I0311 13:39:11.556495 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ec59e42-a1e4-4190-95bc-8370d92be5ff-bundle" (OuterVolumeSpecName: "bundle") pod "5ec59e42-a1e4-4190-95bc-8370d92be5ff" (UID: "5ec59e42-a1e4-4190-95bc-8370d92be5ff"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:39:11 crc kubenswrapper[4923]: I0311 13:39:11.565510 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ec59e42-a1e4-4190-95bc-8370d92be5ff-kube-api-access-mn2qd" (OuterVolumeSpecName: "kube-api-access-mn2qd") pod "5ec59e42-a1e4-4190-95bc-8370d92be5ff" (UID: "5ec59e42-a1e4-4190-95bc-8370d92be5ff"). InnerVolumeSpecName "kube-api-access-mn2qd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:39:11 crc kubenswrapper[4923]: I0311 13:39:11.570852 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ec59e42-a1e4-4190-95bc-8370d92be5ff-util" (OuterVolumeSpecName: "util") pod "5ec59e42-a1e4-4190-95bc-8370d92be5ff" (UID: "5ec59e42-a1e4-4190-95bc-8370d92be5ff"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:39:11 crc kubenswrapper[4923]: I0311 13:39:11.655860 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mn2qd\" (UniqueName: \"kubernetes.io/projected/5ec59e42-a1e4-4190-95bc-8370d92be5ff-kube-api-access-mn2qd\") on node \"crc\" DevicePath \"\"" Mar 11 13:39:11 crc kubenswrapper[4923]: I0311 13:39:11.655907 4923 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5ec59e42-a1e4-4190-95bc-8370d92be5ff-util\") on node \"crc\" DevicePath \"\"" Mar 11 13:39:11 crc kubenswrapper[4923]: I0311 13:39:11.655957 4923 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5ec59e42-a1e4-4190-95bc-8370d92be5ff-bundle\") on node \"crc\" DevicePath \"\"" Mar 11 13:39:12 crc kubenswrapper[4923]: I0311 13:39:12.212863 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf" Mar 11 13:39:12 crc kubenswrapper[4923]: I0311 13:39:12.214734 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf" event={"ID":"5ec59e42-a1e4-4190-95bc-8370d92be5ff","Type":"ContainerDied","Data":"61a271f38be773e1ec078a77fb52279b7c50836e34938ee471cf619027261b7e"} Mar 11 13:39:12 crc kubenswrapper[4923]: I0311 13:39:12.214839 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61a271f38be773e1ec078a77fb52279b7c50836e34938ee471cf619027261b7e" Mar 11 13:39:12 crc kubenswrapper[4923]: I0311 13:39:12.591895 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-sync-cjv6x" Mar 11 13:39:12 crc kubenswrapper[4923]: I0311 13:39:12.676654 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eae8474f-7523-4ee9-8f02-7be328688043-config-data\") pod \"eae8474f-7523-4ee9-8f02-7be328688043\" (UID: \"eae8474f-7523-4ee9-8f02-7be328688043\") " Mar 11 13:39:12 crc kubenswrapper[4923]: I0311 13:39:12.676775 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fq5zx\" (UniqueName: \"kubernetes.io/projected/eae8474f-7523-4ee9-8f02-7be328688043-kube-api-access-fq5zx\") pod \"eae8474f-7523-4ee9-8f02-7be328688043\" (UID: \"eae8474f-7523-4ee9-8f02-7be328688043\") " Mar 11 13:39:12 crc kubenswrapper[4923]: I0311 13:39:12.681226 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eae8474f-7523-4ee9-8f02-7be328688043-kube-api-access-fq5zx" (OuterVolumeSpecName: "kube-api-access-fq5zx") pod "eae8474f-7523-4ee9-8f02-7be328688043" (UID: "eae8474f-7523-4ee9-8f02-7be328688043"). InnerVolumeSpecName "kube-api-access-fq5zx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:39:12 crc kubenswrapper[4923]: I0311 13:39:12.757567 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eae8474f-7523-4ee9-8f02-7be328688043-config-data" (OuterVolumeSpecName: "config-data") pod "eae8474f-7523-4ee9-8f02-7be328688043" (UID: "eae8474f-7523-4ee9-8f02-7be328688043"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:39:12 crc kubenswrapper[4923]: I0311 13:39:12.778736 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eae8474f-7523-4ee9-8f02-7be328688043-config-data\") on node \"crc\" DevicePath \"\"" Mar 11 13:39:12 crc kubenswrapper[4923]: I0311 13:39:12.778789 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fq5zx\" (UniqueName: \"kubernetes.io/projected/eae8474f-7523-4ee9-8f02-7be328688043-kube-api-access-fq5zx\") on node \"crc\" DevicePath \"\"" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.223036 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-db-sync-cjv6x" event={"ID":"eae8474f-7523-4ee9-8f02-7be328688043","Type":"ContainerDied","Data":"6dda9fe864106188b3349b257fb6022c9c2f5495fea427307b1f469904617cf7"} Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.223096 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6dda9fe864106188b3349b257fb6022c9c2f5495fea427307b1f469904617cf7" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.223177 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-db-sync-cjv6x" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.484853 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-nk5mc"] Mar 11 13:39:13 crc kubenswrapper[4923]: E0311 13:39:13.485164 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ec59e42-a1e4-4190-95bc-8370d92be5ff" containerName="util" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.485200 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ec59e42-a1e4-4190-95bc-8370d92be5ff" containerName="util" Mar 11 13:39:13 crc kubenswrapper[4923]: E0311 13:39:13.485220 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ec59e42-a1e4-4190-95bc-8370d92be5ff" containerName="pull" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.485228 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ec59e42-a1e4-4190-95bc-8370d92be5ff" containerName="pull" Mar 11 13:39:13 crc kubenswrapper[4923]: E0311 13:39:13.485245 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1df5f1d-635a-498d-8310-5c8a67181bea" containerName="registry-server" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.485257 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1df5f1d-635a-498d-8310-5c8a67181bea" containerName="registry-server" Mar 11 13:39:13 crc kubenswrapper[4923]: E0311 13:39:13.485266 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eae8474f-7523-4ee9-8f02-7be328688043" containerName="keystone-db-sync" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.485274 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="eae8474f-7523-4ee9-8f02-7be328688043" containerName="keystone-db-sync" Mar 11 13:39:13 crc kubenswrapper[4923]: E0311 13:39:13.485290 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ec59e42-a1e4-4190-95bc-8370d92be5ff" containerName="extract" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.485297 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ec59e42-a1e4-4190-95bc-8370d92be5ff" containerName="extract" Mar 11 13:39:13 crc kubenswrapper[4923]: E0311 13:39:13.485311 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1df5f1d-635a-498d-8310-5c8a67181bea" containerName="extract-content" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.485318 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1df5f1d-635a-498d-8310-5c8a67181bea" containerName="extract-content" Mar 11 13:39:13 crc kubenswrapper[4923]: E0311 13:39:13.485326 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1df5f1d-635a-498d-8310-5c8a67181bea" containerName="extract-utilities" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.485334 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1df5f1d-635a-498d-8310-5c8a67181bea" containerName="extract-utilities" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.485478 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="eae8474f-7523-4ee9-8f02-7be328688043" containerName="keystone-db-sync" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.485499 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ec59e42-a1e4-4190-95bc-8370d92be5ff" containerName="extract" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.485515 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1df5f1d-635a-498d-8310-5c8a67181bea" containerName="registry-server" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.485986 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.492040 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-keystone-dockercfg-7wv7t" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.492254 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-scripts" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.492321 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"osp-secret" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.492567 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.492778 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-config-data" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.505402 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-nk5mc"] Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.601712 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-credential-keys\") pod \"keystone-bootstrap-nk5mc\" (UID: \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\") " pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.601770 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs7wc\" (UniqueName: \"kubernetes.io/projected/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-kube-api-access-bs7wc\") pod \"keystone-bootstrap-nk5mc\" (UID: \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\") " pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.601800 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-fernet-keys\") pod \"keystone-bootstrap-nk5mc\" (UID: \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\") " pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.601834 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-config-data\") pod \"keystone-bootstrap-nk5mc\" (UID: \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\") " pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.601966 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-scripts\") pod \"keystone-bootstrap-nk5mc\" (UID: \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\") " pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.703009 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-scripts\") pod \"keystone-bootstrap-nk5mc\" (UID: \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\") " pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.703098 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-credential-keys\") pod \"keystone-bootstrap-nk5mc\" (UID: \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\") " pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.703144 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs7wc\" (UniqueName: \"kubernetes.io/projected/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-kube-api-access-bs7wc\") pod \"keystone-bootstrap-nk5mc\" (UID: \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\") " pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.703175 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-fernet-keys\") pod \"keystone-bootstrap-nk5mc\" (UID: \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\") " pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.703226 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-config-data\") pod \"keystone-bootstrap-nk5mc\" (UID: \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\") " pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.707094 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-config-data\") pod \"keystone-bootstrap-nk5mc\" (UID: \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\") " pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.707853 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-credential-keys\") pod \"keystone-bootstrap-nk5mc\" (UID: \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\") " pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.709600 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-fernet-keys\") pod \"keystone-bootstrap-nk5mc\" (UID: \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\") " pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.717297 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs7wc\" (UniqueName: \"kubernetes.io/projected/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-kube-api-access-bs7wc\") pod \"keystone-bootstrap-nk5mc\" (UID: \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\") " pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.718470 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-scripts\") pod \"keystone-bootstrap-nk5mc\" (UID: \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\") " pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" Mar 11 13:39:13 crc kubenswrapper[4923]: I0311 13:39:13.835854 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" Mar 11 13:39:14 crc kubenswrapper[4923]: I0311 13:39:14.284473 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-nk5mc"] Mar 11 13:39:15 crc kubenswrapper[4923]: I0311 13:39:15.243659 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" event={"ID":"7ad98c9d-3330-47ad-bf1e-a208ad0401ce","Type":"ContainerStarted","Data":"3a11bd67548b38812031446c2b46e57b76c585f3ad01bfe43de580389986f5a9"} Mar 11 13:39:15 crc kubenswrapper[4923]: I0311 13:39:15.244227 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" event={"ID":"7ad98c9d-3330-47ad-bf1e-a208ad0401ce","Type":"ContainerStarted","Data":"17637585d6ddb311f7b0a78ba80570fe66ece784e733529e98b144452f31d828"} Mar 11 13:39:15 crc kubenswrapper[4923]: I0311 13:39:15.267303 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" podStartSLOduration=2.267271314 podStartE2EDuration="2.267271314s" podCreationTimestamp="2026-03-11 13:39:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:39:15.262849984 +0000 UTC m=+1002.944751788" watchObservedRunningTime="2026-03-11 13:39:15.267271314 +0000 UTC m=+1002.949173068" Mar 11 13:39:17 crc kubenswrapper[4923]: I0311 13:39:17.265953 4923 generic.go:334] "Generic (PLEG): container finished" podID="7ad98c9d-3330-47ad-bf1e-a208ad0401ce" containerID="3a11bd67548b38812031446c2b46e57b76c585f3ad01bfe43de580389986f5a9" exitCode=0 Mar 11 13:39:17 crc kubenswrapper[4923]: I0311 13:39:17.266051 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" event={"ID":"7ad98c9d-3330-47ad-bf1e-a208ad0401ce","Type":"ContainerDied","Data":"3a11bd67548b38812031446c2b46e57b76c585f3ad01bfe43de580389986f5a9"} Mar 11 13:39:18 crc kubenswrapper[4923]: I0311 13:39:18.649825 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" Mar 11 13:39:18 crc kubenswrapper[4923]: I0311 13:39:18.799694 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-fernet-keys\") pod \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\" (UID: \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\") " Mar 11 13:39:18 crc kubenswrapper[4923]: I0311 13:39:18.799798 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-config-data\") pod \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\" (UID: \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\") " Mar 11 13:39:18 crc kubenswrapper[4923]: I0311 13:39:18.799973 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bs7wc\" (UniqueName: \"kubernetes.io/projected/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-kube-api-access-bs7wc\") pod \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\" (UID: \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\") " Mar 11 13:39:18 crc kubenswrapper[4923]: I0311 13:39:18.800116 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-credential-keys\") pod \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\" (UID: \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\") " Mar 11 13:39:18 crc kubenswrapper[4923]: I0311 13:39:18.800394 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-scripts\") pod \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\" (UID: \"7ad98c9d-3330-47ad-bf1e-a208ad0401ce\") " Mar 11 13:39:18 crc kubenswrapper[4923]: I0311 13:39:18.815529 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "7ad98c9d-3330-47ad-bf1e-a208ad0401ce" (UID: "7ad98c9d-3330-47ad-bf1e-a208ad0401ce"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:39:18 crc kubenswrapper[4923]: I0311 13:39:18.819877 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-scripts" (OuterVolumeSpecName: "scripts") pod "7ad98c9d-3330-47ad-bf1e-a208ad0401ce" (UID: "7ad98c9d-3330-47ad-bf1e-a208ad0401ce"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:39:18 crc kubenswrapper[4923]: I0311 13:39:18.822265 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "7ad98c9d-3330-47ad-bf1e-a208ad0401ce" (UID: "7ad98c9d-3330-47ad-bf1e-a208ad0401ce"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:39:18 crc kubenswrapper[4923]: I0311 13:39:18.831224 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-kube-api-access-bs7wc" (OuterVolumeSpecName: "kube-api-access-bs7wc") pod "7ad98c9d-3330-47ad-bf1e-a208ad0401ce" (UID: "7ad98c9d-3330-47ad-bf1e-a208ad0401ce"). InnerVolumeSpecName "kube-api-access-bs7wc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:39:18 crc kubenswrapper[4923]: I0311 13:39:18.834011 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-config-data" (OuterVolumeSpecName: "config-data") pod "7ad98c9d-3330-47ad-bf1e-a208ad0401ce" (UID: "7ad98c9d-3330-47ad-bf1e-a208ad0401ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:39:18 crc kubenswrapper[4923]: I0311 13:39:18.901907 4923 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 11 13:39:18 crc kubenswrapper[4923]: I0311 13:39:18.901940 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-scripts\") on node \"crc\" DevicePath \"\"" Mar 11 13:39:18 crc kubenswrapper[4923]: I0311 13:39:18.901955 4923 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 11 13:39:18 crc kubenswrapper[4923]: I0311 13:39:18.901972 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-config-data\") on node \"crc\" DevicePath \"\"" Mar 11 13:39:18 crc kubenswrapper[4923]: I0311 13:39:18.901984 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bs7wc\" (UniqueName: \"kubernetes.io/projected/7ad98c9d-3330-47ad-bf1e-a208ad0401ce-kube-api-access-bs7wc\") on node \"crc\" DevicePath \"\"" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.287151 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" event={"ID":"7ad98c9d-3330-47ad-bf1e-a208ad0401ce","Type":"ContainerDied","Data":"17637585d6ddb311f7b0a78ba80570fe66ece784e733529e98b144452f31d828"} Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.287198 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17637585d6ddb311f7b0a78ba80570fe66ece784e733529e98b144452f31d828" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.287262 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-bootstrap-nk5mc" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.431131 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/keystone-6b4558db79-pn5px"] Mar 11 13:39:19 crc kubenswrapper[4923]: E0311 13:39:19.431419 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ad98c9d-3330-47ad-bf1e-a208ad0401ce" containerName="keystone-bootstrap" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.431434 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ad98c9d-3330-47ad-bf1e-a208ad0401ce" containerName="keystone-bootstrap" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.431556 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ad98c9d-3330-47ad-bf1e-a208ad0401ce" containerName="keystone-bootstrap" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.431978 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-6b4558db79-pn5px" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.438295 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-keystone-dockercfg-7wv7t" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.438329 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-config-data" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.439075 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.439227 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"keystone-scripts" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.460982 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-6b4558db79-pn5px"] Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.511219 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bda78896-5134-4b9b-917d-40801488fcaa-credential-keys\") pod \"keystone-6b4558db79-pn5px\" (UID: \"bda78896-5134-4b9b-917d-40801488fcaa\") " pod="cinder-kuttl-tests/keystone-6b4558db79-pn5px" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.511315 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bda78896-5134-4b9b-917d-40801488fcaa-scripts\") pod \"keystone-6b4558db79-pn5px\" (UID: \"bda78896-5134-4b9b-917d-40801488fcaa\") " pod="cinder-kuttl-tests/keystone-6b4558db79-pn5px" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.511369 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgzj4\" (UniqueName: \"kubernetes.io/projected/bda78896-5134-4b9b-917d-40801488fcaa-kube-api-access-pgzj4\") pod \"keystone-6b4558db79-pn5px\" (UID: \"bda78896-5134-4b9b-917d-40801488fcaa\") " pod="cinder-kuttl-tests/keystone-6b4558db79-pn5px" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.511401 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bda78896-5134-4b9b-917d-40801488fcaa-config-data\") pod \"keystone-6b4558db79-pn5px\" (UID: \"bda78896-5134-4b9b-917d-40801488fcaa\") " pod="cinder-kuttl-tests/keystone-6b4558db79-pn5px" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.511449 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bda78896-5134-4b9b-917d-40801488fcaa-fernet-keys\") pod \"keystone-6b4558db79-pn5px\" (UID: \"bda78896-5134-4b9b-917d-40801488fcaa\") " pod="cinder-kuttl-tests/keystone-6b4558db79-pn5px" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.613103 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bda78896-5134-4b9b-917d-40801488fcaa-credential-keys\") pod \"keystone-6b4558db79-pn5px\" (UID: \"bda78896-5134-4b9b-917d-40801488fcaa\") " pod="cinder-kuttl-tests/keystone-6b4558db79-pn5px" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.613209 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bda78896-5134-4b9b-917d-40801488fcaa-scripts\") pod \"keystone-6b4558db79-pn5px\" (UID: \"bda78896-5134-4b9b-917d-40801488fcaa\") " pod="cinder-kuttl-tests/keystone-6b4558db79-pn5px" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.613253 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgzj4\" (UniqueName: \"kubernetes.io/projected/bda78896-5134-4b9b-917d-40801488fcaa-kube-api-access-pgzj4\") pod \"keystone-6b4558db79-pn5px\" (UID: \"bda78896-5134-4b9b-917d-40801488fcaa\") " pod="cinder-kuttl-tests/keystone-6b4558db79-pn5px" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.613281 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bda78896-5134-4b9b-917d-40801488fcaa-config-data\") pod \"keystone-6b4558db79-pn5px\" (UID: \"bda78896-5134-4b9b-917d-40801488fcaa\") " pod="cinder-kuttl-tests/keystone-6b4558db79-pn5px" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.613329 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bda78896-5134-4b9b-917d-40801488fcaa-fernet-keys\") pod \"keystone-6b4558db79-pn5px\" (UID: \"bda78896-5134-4b9b-917d-40801488fcaa\") " pod="cinder-kuttl-tests/keystone-6b4558db79-pn5px" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.619386 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bda78896-5134-4b9b-917d-40801488fcaa-fernet-keys\") pod \"keystone-6b4558db79-pn5px\" (UID: \"bda78896-5134-4b9b-917d-40801488fcaa\") " pod="cinder-kuttl-tests/keystone-6b4558db79-pn5px" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.621806 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bda78896-5134-4b9b-917d-40801488fcaa-credential-keys\") pod \"keystone-6b4558db79-pn5px\" (UID: \"bda78896-5134-4b9b-917d-40801488fcaa\") " pod="cinder-kuttl-tests/keystone-6b4558db79-pn5px" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.625050 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bda78896-5134-4b9b-917d-40801488fcaa-config-data\") pod \"keystone-6b4558db79-pn5px\" (UID: \"bda78896-5134-4b9b-917d-40801488fcaa\") " pod="cinder-kuttl-tests/keystone-6b4558db79-pn5px" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.625333 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bda78896-5134-4b9b-917d-40801488fcaa-scripts\") pod \"keystone-6b4558db79-pn5px\" (UID: \"bda78896-5134-4b9b-917d-40801488fcaa\") " pod="cinder-kuttl-tests/keystone-6b4558db79-pn5px" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.647423 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgzj4\" (UniqueName: \"kubernetes.io/projected/bda78896-5134-4b9b-917d-40801488fcaa-kube-api-access-pgzj4\") pod \"keystone-6b4558db79-pn5px\" (UID: \"bda78896-5134-4b9b-917d-40801488fcaa\") " pod="cinder-kuttl-tests/keystone-6b4558db79-pn5px" Mar 11 13:39:19 crc kubenswrapper[4923]: I0311 13:39:19.752719 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/keystone-6b4558db79-pn5px" Mar 11 13:39:20 crc kubenswrapper[4923]: I0311 13:39:20.268650 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/keystone-6b4558db79-pn5px"] Mar 11 13:39:20 crc kubenswrapper[4923]: I0311 13:39:20.294050 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-6b4558db79-pn5px" event={"ID":"bda78896-5134-4b9b-917d-40801488fcaa","Type":"ContainerStarted","Data":"62d4fbe68ce07275d9e1a849b3b33746f6bda57947b4b65949afdec345860eed"} Mar 11 13:39:21 crc kubenswrapper[4923]: I0311 13:39:21.302918 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/keystone-6b4558db79-pn5px" event={"ID":"bda78896-5134-4b9b-917d-40801488fcaa","Type":"ContainerStarted","Data":"f8a429d831761fd5a3572cb33eb3e66aed0caca9e65ea793a9bdd8279350a72a"} Mar 11 13:39:21 crc kubenswrapper[4923]: I0311 13:39:21.303392 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/keystone-6b4558db79-pn5px" Mar 11 13:39:21 crc kubenswrapper[4923]: I0311 13:39:21.331508 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/keystone-6b4558db79-pn5px" podStartSLOduration=2.331473706 podStartE2EDuration="2.331473706s" podCreationTimestamp="2026-03-11 13:39:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:39:21.323329063 +0000 UTC m=+1009.005230787" watchObservedRunningTime="2026-03-11 13:39:21.331473706 +0000 UTC m=+1009.013375480" Mar 11 13:39:24 crc kubenswrapper[4923]: I0311 13:39:24.021750 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8698b9b776-6c84l"] Mar 11 13:39:24 crc kubenswrapper[4923]: I0311 13:39:24.023820 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8698b9b776-6c84l" Mar 11 13:39:24 crc kubenswrapper[4923]: I0311 13:39:24.026179 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-wgqps" Mar 11 13:39:24 crc kubenswrapper[4923]: I0311 13:39:24.026187 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-service-cert" Mar 11 13:39:24 crc kubenswrapper[4923]: I0311 13:39:24.051767 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8698b9b776-6c84l"] Mar 11 13:39:24 crc kubenswrapper[4923]: I0311 13:39:24.179927 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b2d6a790-42f2-4c33-a7d5-24e033961c36-webhook-cert\") pod \"cinder-operator-controller-manager-8698b9b776-6c84l\" (UID: \"b2d6a790-42f2-4c33-a7d5-24e033961c36\") " pod="openstack-operators/cinder-operator-controller-manager-8698b9b776-6c84l" Mar 11 13:39:24 crc kubenswrapper[4923]: I0311 13:39:24.180067 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b2d6a790-42f2-4c33-a7d5-24e033961c36-apiservice-cert\") pod \"cinder-operator-controller-manager-8698b9b776-6c84l\" (UID: \"b2d6a790-42f2-4c33-a7d5-24e033961c36\") " pod="openstack-operators/cinder-operator-controller-manager-8698b9b776-6c84l" Mar 11 13:39:24 crc kubenswrapper[4923]: I0311 13:39:24.180101 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2k5j6\" (UniqueName: \"kubernetes.io/projected/b2d6a790-42f2-4c33-a7d5-24e033961c36-kube-api-access-2k5j6\") pod \"cinder-operator-controller-manager-8698b9b776-6c84l\" (UID: \"b2d6a790-42f2-4c33-a7d5-24e033961c36\") " pod="openstack-operators/cinder-operator-controller-manager-8698b9b776-6c84l" Mar 11 13:39:24 crc kubenswrapper[4923]: I0311 13:39:24.281598 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b2d6a790-42f2-4c33-a7d5-24e033961c36-webhook-cert\") pod \"cinder-operator-controller-manager-8698b9b776-6c84l\" (UID: \"b2d6a790-42f2-4c33-a7d5-24e033961c36\") " pod="openstack-operators/cinder-operator-controller-manager-8698b9b776-6c84l" Mar 11 13:39:24 crc kubenswrapper[4923]: I0311 13:39:24.281798 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b2d6a790-42f2-4c33-a7d5-24e033961c36-apiservice-cert\") pod \"cinder-operator-controller-manager-8698b9b776-6c84l\" (UID: \"b2d6a790-42f2-4c33-a7d5-24e033961c36\") " pod="openstack-operators/cinder-operator-controller-manager-8698b9b776-6c84l" Mar 11 13:39:24 crc kubenswrapper[4923]: I0311 13:39:24.281856 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2k5j6\" (UniqueName: \"kubernetes.io/projected/b2d6a790-42f2-4c33-a7d5-24e033961c36-kube-api-access-2k5j6\") pod \"cinder-operator-controller-manager-8698b9b776-6c84l\" (UID: \"b2d6a790-42f2-4c33-a7d5-24e033961c36\") " pod="openstack-operators/cinder-operator-controller-manager-8698b9b776-6c84l" Mar 11 13:39:24 crc kubenswrapper[4923]: I0311 13:39:24.289205 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b2d6a790-42f2-4c33-a7d5-24e033961c36-apiservice-cert\") pod \"cinder-operator-controller-manager-8698b9b776-6c84l\" (UID: \"b2d6a790-42f2-4c33-a7d5-24e033961c36\") " pod="openstack-operators/cinder-operator-controller-manager-8698b9b776-6c84l" Mar 11 13:39:24 crc kubenswrapper[4923]: I0311 13:39:24.289205 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b2d6a790-42f2-4c33-a7d5-24e033961c36-webhook-cert\") pod \"cinder-operator-controller-manager-8698b9b776-6c84l\" (UID: \"b2d6a790-42f2-4c33-a7d5-24e033961c36\") " pod="openstack-operators/cinder-operator-controller-manager-8698b9b776-6c84l" Mar 11 13:39:24 crc kubenswrapper[4923]: I0311 13:39:24.299625 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2k5j6\" (UniqueName: \"kubernetes.io/projected/b2d6a790-42f2-4c33-a7d5-24e033961c36-kube-api-access-2k5j6\") pod \"cinder-operator-controller-manager-8698b9b776-6c84l\" (UID: \"b2d6a790-42f2-4c33-a7d5-24e033961c36\") " pod="openstack-operators/cinder-operator-controller-manager-8698b9b776-6c84l" Mar 11 13:39:24 crc kubenswrapper[4923]: I0311 13:39:24.343671 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8698b9b776-6c84l" Mar 11 13:39:24 crc kubenswrapper[4923]: I0311 13:39:24.762455 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8698b9b776-6c84l"] Mar 11 13:39:24 crc kubenswrapper[4923]: W0311 13:39:24.768614 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2d6a790_42f2_4c33_a7d5_24e033961c36.slice/crio-69bde22e1922bb31c79ece71e5d858fe4a8670964289bac7b526b3d45fd81f13 WatchSource:0}: Error finding container 69bde22e1922bb31c79ece71e5d858fe4a8670964289bac7b526b3d45fd81f13: Status 404 returned error can't find the container with id 69bde22e1922bb31c79ece71e5d858fe4a8670964289bac7b526b3d45fd81f13 Mar 11 13:39:25 crc kubenswrapper[4923]: I0311 13:39:25.338583 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8698b9b776-6c84l" event={"ID":"b2d6a790-42f2-4c33-a7d5-24e033961c36","Type":"ContainerStarted","Data":"69bde22e1922bb31c79ece71e5d858fe4a8670964289bac7b526b3d45fd81f13"} Mar 11 13:39:27 crc kubenswrapper[4923]: I0311 13:39:27.352591 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8698b9b776-6c84l" event={"ID":"b2d6a790-42f2-4c33-a7d5-24e033961c36","Type":"ContainerStarted","Data":"708db36fcf4730d6b6676f99664380dcad6bf5e830512722e4b5e8ad19735a15"} Mar 11 13:39:27 crc kubenswrapper[4923]: I0311 13:39:27.353097 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-8698b9b776-6c84l" Mar 11 13:39:27 crc kubenswrapper[4923]: I0311 13:39:27.379275 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-8698b9b776-6c84l" podStartSLOduration=1.937514016 podStartE2EDuration="3.379242119s" podCreationTimestamp="2026-03-11 13:39:24 +0000 UTC" firstStartedPulling="2026-03-11 13:39:24.770698571 +0000 UTC m=+1012.452600285" lastFinishedPulling="2026-03-11 13:39:26.212426674 +0000 UTC m=+1013.894328388" observedRunningTime="2026-03-11 13:39:27.370735407 +0000 UTC m=+1015.052637121" watchObservedRunningTime="2026-03-11 13:39:27.379242119 +0000 UTC m=+1015.061143873" Mar 11 13:39:34 crc kubenswrapper[4923]: I0311 13:39:34.349581 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-8698b9b776-6c84l" Mar 11 13:39:48 crc kubenswrapper[4923]: I0311 13:39:48.669676 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-e6fe-account-create-update-w2s4t"] Mar 11 13:39:48 crc kubenswrapper[4923]: I0311 13:39:48.671126 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-e6fe-account-create-update-w2s4t" Mar 11 13:39:48 crc kubenswrapper[4923]: I0311 13:39:48.688401 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-db-create-24dq4"] Mar 11 13:39:48 crc kubenswrapper[4923]: I0311 13:39:48.689233 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-24dq4"] Mar 11 13:39:48 crc kubenswrapper[4923]: I0311 13:39:48.693453 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-e6fe-account-create-update-w2s4t"] Mar 11 13:39:48 crc kubenswrapper[4923]: I0311 13:39:48.693553 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-24dq4" Mar 11 13:39:48 crc kubenswrapper[4923]: I0311 13:39:48.702980 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-db-secret" Mar 11 13:39:48 crc kubenswrapper[4923]: I0311 13:39:48.733158 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6-operator-scripts\") pod \"cinder-db-create-24dq4\" (UID: \"1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6\") " pod="cinder-kuttl-tests/cinder-db-create-24dq4" Mar 11 13:39:48 crc kubenswrapper[4923]: I0311 13:39:48.733210 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da36884b-e7cc-4ba1-9966-79b1a14059da-operator-scripts\") pod \"cinder-e6fe-account-create-update-w2s4t\" (UID: \"da36884b-e7cc-4ba1-9966-79b1a14059da\") " pod="cinder-kuttl-tests/cinder-e6fe-account-create-update-w2s4t" Mar 11 13:39:48 crc kubenswrapper[4923]: I0311 13:39:48.733262 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljzft\" (UniqueName: \"kubernetes.io/projected/da36884b-e7cc-4ba1-9966-79b1a14059da-kube-api-access-ljzft\") pod \"cinder-e6fe-account-create-update-w2s4t\" (UID: \"da36884b-e7cc-4ba1-9966-79b1a14059da\") " pod="cinder-kuttl-tests/cinder-e6fe-account-create-update-w2s4t" Mar 11 13:39:48 crc kubenswrapper[4923]: I0311 13:39:48.733299 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzvmc\" (UniqueName: \"kubernetes.io/projected/1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6-kube-api-access-qzvmc\") pod \"cinder-db-create-24dq4\" (UID: \"1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6\") " pod="cinder-kuttl-tests/cinder-db-create-24dq4" Mar 11 13:39:48 crc kubenswrapper[4923]: I0311 13:39:48.834830 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6-operator-scripts\") pod \"cinder-db-create-24dq4\" (UID: \"1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6\") " pod="cinder-kuttl-tests/cinder-db-create-24dq4" Mar 11 13:39:48 crc kubenswrapper[4923]: I0311 13:39:48.834889 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da36884b-e7cc-4ba1-9966-79b1a14059da-operator-scripts\") pod \"cinder-e6fe-account-create-update-w2s4t\" (UID: \"da36884b-e7cc-4ba1-9966-79b1a14059da\") " pod="cinder-kuttl-tests/cinder-e6fe-account-create-update-w2s4t" Mar 11 13:39:48 crc kubenswrapper[4923]: I0311 13:39:48.834933 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljzft\" (UniqueName: \"kubernetes.io/projected/da36884b-e7cc-4ba1-9966-79b1a14059da-kube-api-access-ljzft\") pod \"cinder-e6fe-account-create-update-w2s4t\" (UID: \"da36884b-e7cc-4ba1-9966-79b1a14059da\") " pod="cinder-kuttl-tests/cinder-e6fe-account-create-update-w2s4t" Mar 11 13:39:48 crc kubenswrapper[4923]: I0311 13:39:48.834963 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzvmc\" (UniqueName: \"kubernetes.io/projected/1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6-kube-api-access-qzvmc\") pod \"cinder-db-create-24dq4\" (UID: \"1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6\") " pod="cinder-kuttl-tests/cinder-db-create-24dq4" Mar 11 13:39:48 crc kubenswrapper[4923]: I0311 13:39:48.835798 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6-operator-scripts\") pod \"cinder-db-create-24dq4\" (UID: \"1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6\") " pod="cinder-kuttl-tests/cinder-db-create-24dq4" Mar 11 13:39:48 crc kubenswrapper[4923]: I0311 13:39:48.836870 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da36884b-e7cc-4ba1-9966-79b1a14059da-operator-scripts\") pod \"cinder-e6fe-account-create-update-w2s4t\" (UID: \"da36884b-e7cc-4ba1-9966-79b1a14059da\") " pod="cinder-kuttl-tests/cinder-e6fe-account-create-update-w2s4t" Mar 11 13:39:48 crc kubenswrapper[4923]: I0311 13:39:48.853850 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzvmc\" (UniqueName: \"kubernetes.io/projected/1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6-kube-api-access-qzvmc\") pod \"cinder-db-create-24dq4\" (UID: \"1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6\") " pod="cinder-kuttl-tests/cinder-db-create-24dq4" Mar 11 13:39:48 crc kubenswrapper[4923]: I0311 13:39:48.854494 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljzft\" (UniqueName: \"kubernetes.io/projected/da36884b-e7cc-4ba1-9966-79b1a14059da-kube-api-access-ljzft\") pod \"cinder-e6fe-account-create-update-w2s4t\" (UID: \"da36884b-e7cc-4ba1-9966-79b1a14059da\") " pod="cinder-kuttl-tests/cinder-e6fe-account-create-update-w2s4t" Mar 11 13:39:49 crc kubenswrapper[4923]: I0311 13:39:49.022515 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-e6fe-account-create-update-w2s4t" Mar 11 13:39:49 crc kubenswrapper[4923]: I0311 13:39:49.033229 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-24dq4" Mar 11 13:39:49 crc kubenswrapper[4923]: I0311 13:39:49.501764 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-e6fe-account-create-update-w2s4t"] Mar 11 13:39:49 crc kubenswrapper[4923]: I0311 13:39:49.527913 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-e6fe-account-create-update-w2s4t" event={"ID":"da36884b-e7cc-4ba1-9966-79b1a14059da","Type":"ContainerStarted","Data":"43f2a4598f8ab1b50cd8a87529f1fb99992f57ff570392df1f01acdd5a7bd0c7"} Mar 11 13:39:49 crc kubenswrapper[4923]: W0311 13:39:49.608562 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b2f3f3b_b87d_4a10_adfb_38cbdb70b2c6.slice/crio-20244e048fb210a9f22384ea6a636184ded4a913d55e18882d819b3488cd09a6 WatchSource:0}: Error finding container 20244e048fb210a9f22384ea6a636184ded4a913d55e18882d819b3488cd09a6: Status 404 returned error can't find the container with id 20244e048fb210a9f22384ea6a636184ded4a913d55e18882d819b3488cd09a6 Mar 11 13:39:49 crc kubenswrapper[4923]: I0311 13:39:49.613399 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-24dq4"] Mar 11 13:39:50 crc kubenswrapper[4923]: I0311 13:39:50.540900 4923 generic.go:334] "Generic (PLEG): container finished" podID="da36884b-e7cc-4ba1-9966-79b1a14059da" containerID="fca8ac0b773cafac5f5ef9ba283429f0847b0ce13ccb570af77cb877754851e2" exitCode=0 Mar 11 13:39:50 crc kubenswrapper[4923]: I0311 13:39:50.541462 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-e6fe-account-create-update-w2s4t" event={"ID":"da36884b-e7cc-4ba1-9966-79b1a14059da","Type":"ContainerDied","Data":"fca8ac0b773cafac5f5ef9ba283429f0847b0ce13ccb570af77cb877754851e2"} Mar 11 13:39:50 crc kubenswrapper[4923]: I0311 13:39:50.553779 4923 generic.go:334] "Generic (PLEG): container finished" podID="1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6" containerID="e437e1c2f3c3617a128fbb5b3bb5b67e8a7d4caa69c51814e865f73459311004" exitCode=0 Mar 11 13:39:50 crc kubenswrapper[4923]: I0311 13:39:50.553844 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-24dq4" event={"ID":"1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6","Type":"ContainerDied","Data":"e437e1c2f3c3617a128fbb5b3bb5b67e8a7d4caa69c51814e865f73459311004"} Mar 11 13:39:50 crc kubenswrapper[4923]: I0311 13:39:50.553881 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-24dq4" event={"ID":"1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6","Type":"ContainerStarted","Data":"20244e048fb210a9f22384ea6a636184ded4a913d55e18882d819b3488cd09a6"} Mar 11 13:39:51 crc kubenswrapper[4923]: I0311 13:39:51.200593 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/keystone-6b4558db79-pn5px" Mar 11 13:39:51 crc kubenswrapper[4923]: I0311 13:39:51.907066 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-e6fe-account-create-update-w2s4t" Mar 11 13:39:51 crc kubenswrapper[4923]: I0311 13:39:51.975549 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-24dq4" Mar 11 13:39:51 crc kubenswrapper[4923]: I0311 13:39:51.981296 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da36884b-e7cc-4ba1-9966-79b1a14059da-operator-scripts\") pod \"da36884b-e7cc-4ba1-9966-79b1a14059da\" (UID: \"da36884b-e7cc-4ba1-9966-79b1a14059da\") " Mar 11 13:39:51 crc kubenswrapper[4923]: I0311 13:39:51.981410 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljzft\" (UniqueName: \"kubernetes.io/projected/da36884b-e7cc-4ba1-9966-79b1a14059da-kube-api-access-ljzft\") pod \"da36884b-e7cc-4ba1-9966-79b1a14059da\" (UID: \"da36884b-e7cc-4ba1-9966-79b1a14059da\") " Mar 11 13:39:51 crc kubenswrapper[4923]: I0311 13:39:51.982389 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da36884b-e7cc-4ba1-9966-79b1a14059da-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "da36884b-e7cc-4ba1-9966-79b1a14059da" (UID: "da36884b-e7cc-4ba1-9966-79b1a14059da"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:39:51 crc kubenswrapper[4923]: I0311 13:39:51.989710 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da36884b-e7cc-4ba1-9966-79b1a14059da-kube-api-access-ljzft" (OuterVolumeSpecName: "kube-api-access-ljzft") pod "da36884b-e7cc-4ba1-9966-79b1a14059da" (UID: "da36884b-e7cc-4ba1-9966-79b1a14059da"). InnerVolumeSpecName "kube-api-access-ljzft". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:39:52 crc kubenswrapper[4923]: I0311 13:39:52.082280 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6-operator-scripts\") pod \"1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6\" (UID: \"1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6\") " Mar 11 13:39:52 crc kubenswrapper[4923]: I0311 13:39:52.082317 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qzvmc\" (UniqueName: \"kubernetes.io/projected/1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6-kube-api-access-qzvmc\") pod \"1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6\" (UID: \"1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6\") " Mar 11 13:39:52 crc kubenswrapper[4923]: I0311 13:39:52.082627 4923 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/da36884b-e7cc-4ba1-9966-79b1a14059da-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 11 13:39:52 crc kubenswrapper[4923]: I0311 13:39:52.082638 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljzft\" (UniqueName: \"kubernetes.io/projected/da36884b-e7cc-4ba1-9966-79b1a14059da-kube-api-access-ljzft\") on node \"crc\" DevicePath \"\"" Mar 11 13:39:52 crc kubenswrapper[4923]: I0311 13:39:52.082756 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6" (UID: "1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:39:52 crc kubenswrapper[4923]: I0311 13:39:52.084955 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6-kube-api-access-qzvmc" (OuterVolumeSpecName: "kube-api-access-qzvmc") pod "1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6" (UID: "1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6"). InnerVolumeSpecName "kube-api-access-qzvmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:39:52 crc kubenswrapper[4923]: I0311 13:39:52.184185 4923 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 11 13:39:52 crc kubenswrapper[4923]: I0311 13:39:52.184214 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qzvmc\" (UniqueName: \"kubernetes.io/projected/1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6-kube-api-access-qzvmc\") on node \"crc\" DevicePath \"\"" Mar 11 13:39:52 crc kubenswrapper[4923]: I0311 13:39:52.576382 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-e6fe-account-create-update-w2s4t" event={"ID":"da36884b-e7cc-4ba1-9966-79b1a14059da","Type":"ContainerDied","Data":"43f2a4598f8ab1b50cd8a87529f1fb99992f57ff570392df1f01acdd5a7bd0c7"} Mar 11 13:39:52 crc kubenswrapper[4923]: I0311 13:39:52.576491 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43f2a4598f8ab1b50cd8a87529f1fb99992f57ff570392df1f01acdd5a7bd0c7" Mar 11 13:39:52 crc kubenswrapper[4923]: I0311 13:39:52.576398 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-e6fe-account-create-update-w2s4t" Mar 11 13:39:52 crc kubenswrapper[4923]: I0311 13:39:52.578900 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-create-24dq4" event={"ID":"1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6","Type":"ContainerDied","Data":"20244e048fb210a9f22384ea6a636184ded4a913d55e18882d819b3488cd09a6"} Mar 11 13:39:52 crc kubenswrapper[4923]: I0311 13:39:52.578945 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20244e048fb210a9f22384ea6a636184ded4a913d55e18882d819b3488cd09a6" Mar 11 13:39:52 crc kubenswrapper[4923]: I0311 13:39:52.579017 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-create-24dq4" Mar 11 13:39:53 crc kubenswrapper[4923]: I0311 13:39:53.929983 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-gbxj9"] Mar 11 13:39:53 crc kubenswrapper[4923]: E0311 13:39:53.931763 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da36884b-e7cc-4ba1-9966-79b1a14059da" containerName="mariadb-account-create-update" Mar 11 13:39:53 crc kubenswrapper[4923]: I0311 13:39:53.931833 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="da36884b-e7cc-4ba1-9966-79b1a14059da" containerName="mariadb-account-create-update" Mar 11 13:39:53 crc kubenswrapper[4923]: E0311 13:39:53.931872 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6" containerName="mariadb-database-create" Mar 11 13:39:53 crc kubenswrapper[4923]: I0311 13:39:53.931888 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6" containerName="mariadb-database-create" Mar 11 13:39:53 crc kubenswrapper[4923]: I0311 13:39:53.932213 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6" containerName="mariadb-database-create" Mar 11 13:39:53 crc kubenswrapper[4923]: I0311 13:39:53.932284 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="da36884b-e7cc-4ba1-9966-79b1a14059da" containerName="mariadb-account-create-update" Mar 11 13:39:53 crc kubenswrapper[4923]: I0311 13:39:53.933216 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" Mar 11 13:39:53 crc kubenswrapper[4923]: I0311 13:39:53.936308 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-config-data" Mar 11 13:39:53 crc kubenswrapper[4923]: I0311 13:39:53.938307 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scripts" Mar 11 13:39:53 crc kubenswrapper[4923]: I0311 13:39:53.944114 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-cinder-dockercfg-pkktf" Mar 11 13:39:53 crc kubenswrapper[4923]: I0311 13:39:53.946118 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-gbxj9"] Mar 11 13:39:54 crc kubenswrapper[4923]: I0311 13:39:54.013421 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f7361c58-e41e-42ab-9c24-d0e470ccea54-etc-machine-id\") pod \"cinder-db-sync-gbxj9\" (UID: \"f7361c58-e41e-42ab-9c24-d0e470ccea54\") " pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" Mar 11 13:39:54 crc kubenswrapper[4923]: I0311 13:39:54.013499 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xg9h7\" (UniqueName: \"kubernetes.io/projected/f7361c58-e41e-42ab-9c24-d0e470ccea54-kube-api-access-xg9h7\") pod \"cinder-db-sync-gbxj9\" (UID: \"f7361c58-e41e-42ab-9c24-d0e470ccea54\") " pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" Mar 11 13:39:54 crc kubenswrapper[4923]: I0311 13:39:54.013552 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f7361c58-e41e-42ab-9c24-d0e470ccea54-db-sync-config-data\") pod \"cinder-db-sync-gbxj9\" (UID: \"f7361c58-e41e-42ab-9c24-d0e470ccea54\") " pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" Mar 11 13:39:54 crc kubenswrapper[4923]: I0311 13:39:54.013625 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7361c58-e41e-42ab-9c24-d0e470ccea54-config-data\") pod \"cinder-db-sync-gbxj9\" (UID: \"f7361c58-e41e-42ab-9c24-d0e470ccea54\") " pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" Mar 11 13:39:54 crc kubenswrapper[4923]: I0311 13:39:54.013777 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7361c58-e41e-42ab-9c24-d0e470ccea54-scripts\") pod \"cinder-db-sync-gbxj9\" (UID: \"f7361c58-e41e-42ab-9c24-d0e470ccea54\") " pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" Mar 11 13:39:54 crc kubenswrapper[4923]: I0311 13:39:54.114305 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f7361c58-e41e-42ab-9c24-d0e470ccea54-etc-machine-id\") pod \"cinder-db-sync-gbxj9\" (UID: \"f7361c58-e41e-42ab-9c24-d0e470ccea54\") " pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" Mar 11 13:39:54 crc kubenswrapper[4923]: I0311 13:39:54.114423 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xg9h7\" (UniqueName: \"kubernetes.io/projected/f7361c58-e41e-42ab-9c24-d0e470ccea54-kube-api-access-xg9h7\") pod \"cinder-db-sync-gbxj9\" (UID: \"f7361c58-e41e-42ab-9c24-d0e470ccea54\") " pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" Mar 11 13:39:54 crc kubenswrapper[4923]: I0311 13:39:54.114433 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f7361c58-e41e-42ab-9c24-d0e470ccea54-etc-machine-id\") pod \"cinder-db-sync-gbxj9\" (UID: \"f7361c58-e41e-42ab-9c24-d0e470ccea54\") " pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" Mar 11 13:39:54 crc kubenswrapper[4923]: I0311 13:39:54.114470 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f7361c58-e41e-42ab-9c24-d0e470ccea54-db-sync-config-data\") pod \"cinder-db-sync-gbxj9\" (UID: \"f7361c58-e41e-42ab-9c24-d0e470ccea54\") " pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" Mar 11 13:39:54 crc kubenswrapper[4923]: I0311 13:39:54.114519 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7361c58-e41e-42ab-9c24-d0e470ccea54-config-data\") pod \"cinder-db-sync-gbxj9\" (UID: \"f7361c58-e41e-42ab-9c24-d0e470ccea54\") " pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" Mar 11 13:39:54 crc kubenswrapper[4923]: I0311 13:39:54.114543 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7361c58-e41e-42ab-9c24-d0e470ccea54-scripts\") pod \"cinder-db-sync-gbxj9\" (UID: \"f7361c58-e41e-42ab-9c24-d0e470ccea54\") " pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" Mar 11 13:39:54 crc kubenswrapper[4923]: I0311 13:39:54.119955 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f7361c58-e41e-42ab-9c24-d0e470ccea54-db-sync-config-data\") pod \"cinder-db-sync-gbxj9\" (UID: \"f7361c58-e41e-42ab-9c24-d0e470ccea54\") " pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" Mar 11 13:39:54 crc kubenswrapper[4923]: I0311 13:39:54.120206 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7361c58-e41e-42ab-9c24-d0e470ccea54-scripts\") pod \"cinder-db-sync-gbxj9\" (UID: \"f7361c58-e41e-42ab-9c24-d0e470ccea54\") " pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" Mar 11 13:39:54 crc kubenswrapper[4923]: I0311 13:39:54.120845 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7361c58-e41e-42ab-9c24-d0e470ccea54-config-data\") pod \"cinder-db-sync-gbxj9\" (UID: \"f7361c58-e41e-42ab-9c24-d0e470ccea54\") " pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" Mar 11 13:39:54 crc kubenswrapper[4923]: I0311 13:39:54.149579 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xg9h7\" (UniqueName: \"kubernetes.io/projected/f7361c58-e41e-42ab-9c24-d0e470ccea54-kube-api-access-xg9h7\") pod \"cinder-db-sync-gbxj9\" (UID: \"f7361c58-e41e-42ab-9c24-d0e470ccea54\") " pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" Mar 11 13:39:54 crc kubenswrapper[4923]: I0311 13:39:54.269253 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" Mar 11 13:39:54 crc kubenswrapper[4923]: I0311 13:39:54.778072 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-gbxj9"] Mar 11 13:39:54 crc kubenswrapper[4923]: W0311 13:39:54.784167 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf7361c58_e41e_42ab_9c24_d0e470ccea54.slice/crio-e8c0a3b8167e29c05e43f023027501f07991fa3eefe0315779b1f69101dbdb72 WatchSource:0}: Error finding container e8c0a3b8167e29c05e43f023027501f07991fa3eefe0315779b1f69101dbdb72: Status 404 returned error can't find the container with id e8c0a3b8167e29c05e43f023027501f07991fa3eefe0315779b1f69101dbdb72 Mar 11 13:39:55 crc kubenswrapper[4923]: I0311 13:39:55.603686 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" event={"ID":"f7361c58-e41e-42ab-9c24-d0e470ccea54","Type":"ContainerStarted","Data":"e8c0a3b8167e29c05e43f023027501f07991fa3eefe0315779b1f69101dbdb72"} Mar 11 13:40:00 crc kubenswrapper[4923]: I0311 13:40:00.127760 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29553940-wj5ml"] Mar 11 13:40:00 crc kubenswrapper[4923]: I0311 13:40:00.132527 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553940-wj5ml" Mar 11 13:40:00 crc kubenswrapper[4923]: I0311 13:40:00.136969 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-nxck7" Mar 11 13:40:00 crc kubenswrapper[4923]: I0311 13:40:00.138018 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 11 13:40:00 crc kubenswrapper[4923]: I0311 13:40:00.138613 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 11 13:40:00 crc kubenswrapper[4923]: I0311 13:40:00.146211 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553940-wj5ml"] Mar 11 13:40:00 crc kubenswrapper[4923]: I0311 13:40:00.210315 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5nj5\" (UniqueName: \"kubernetes.io/projected/13c70bc8-3f75-401b-8a96-d3d044e39672-kube-api-access-t5nj5\") pod \"auto-csr-approver-29553940-wj5ml\" (UID: \"13c70bc8-3f75-401b-8a96-d3d044e39672\") " pod="openshift-infra/auto-csr-approver-29553940-wj5ml" Mar 11 13:40:00 crc kubenswrapper[4923]: I0311 13:40:00.311796 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5nj5\" (UniqueName: \"kubernetes.io/projected/13c70bc8-3f75-401b-8a96-d3d044e39672-kube-api-access-t5nj5\") pod \"auto-csr-approver-29553940-wj5ml\" (UID: \"13c70bc8-3f75-401b-8a96-d3d044e39672\") " pod="openshift-infra/auto-csr-approver-29553940-wj5ml" Mar 11 13:40:00 crc kubenswrapper[4923]: I0311 13:40:00.336314 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5nj5\" (UniqueName: \"kubernetes.io/projected/13c70bc8-3f75-401b-8a96-d3d044e39672-kube-api-access-t5nj5\") pod \"auto-csr-approver-29553940-wj5ml\" (UID: \"13c70bc8-3f75-401b-8a96-d3d044e39672\") " pod="openshift-infra/auto-csr-approver-29553940-wj5ml" Mar 11 13:40:00 crc kubenswrapper[4923]: I0311 13:40:00.464751 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553940-wj5ml" Mar 11 13:40:09 crc kubenswrapper[4923]: I0311 13:40:09.258044 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553940-wj5ml"] Mar 11 13:40:09 crc kubenswrapper[4923]: I0311 13:40:09.702257 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553940-wj5ml" event={"ID":"13c70bc8-3f75-401b-8a96-d3d044e39672","Type":"ContainerStarted","Data":"5b17fc1e9108d770cef387d2219540ae54822e0433148406be64e168b4ac9f04"} Mar 11 13:40:10 crc kubenswrapper[4923]: I0311 13:40:10.722673 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" event={"ID":"f7361c58-e41e-42ab-9c24-d0e470ccea54","Type":"ContainerStarted","Data":"304bb4d25b0ee4a7e44d56d37696f9762d3859438b560ce2b1936ba1a46dd358"} Mar 11 13:40:10 crc kubenswrapper[4923]: I0311 13:40:10.724762 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553940-wj5ml" event={"ID":"13c70bc8-3f75-401b-8a96-d3d044e39672","Type":"ContainerStarted","Data":"fb6ab8fb89f37fb010e2ca58da098bf09d4aeb72cf5ee3378de079317faf467c"} Mar 11 13:40:10 crc kubenswrapper[4923]: I0311 13:40:10.749560 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" podStartSLOduration=3.37996214 podStartE2EDuration="17.749524722s" podCreationTimestamp="2026-03-11 13:39:53 +0000 UTC" firstStartedPulling="2026-03-11 13:39:54.793038905 +0000 UTC m=+1042.474940619" lastFinishedPulling="2026-03-11 13:40:09.162601487 +0000 UTC m=+1056.844503201" observedRunningTime="2026-03-11 13:40:10.741597272 +0000 UTC m=+1058.423499026" watchObservedRunningTime="2026-03-11 13:40:10.749524722 +0000 UTC m=+1058.431426446" Mar 11 13:40:10 crc kubenswrapper[4923]: I0311 13:40:10.775305 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29553940-wj5ml" podStartSLOduration=9.870494722 podStartE2EDuration="10.775277215s" podCreationTimestamp="2026-03-11 13:40:00 +0000 UTC" firstStartedPulling="2026-03-11 13:40:09.267524575 +0000 UTC m=+1056.949426289" lastFinishedPulling="2026-03-11 13:40:10.172307078 +0000 UTC m=+1057.854208782" observedRunningTime="2026-03-11 13:40:10.769826687 +0000 UTC m=+1058.451728411" watchObservedRunningTime="2026-03-11 13:40:10.775277215 +0000 UTC m=+1058.457178939" Mar 11 13:40:11 crc kubenswrapper[4923]: I0311 13:40:11.736028 4923 generic.go:334] "Generic (PLEG): container finished" podID="13c70bc8-3f75-401b-8a96-d3d044e39672" containerID="fb6ab8fb89f37fb010e2ca58da098bf09d4aeb72cf5ee3378de079317faf467c" exitCode=0 Mar 11 13:40:11 crc kubenswrapper[4923]: I0311 13:40:11.736097 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553940-wj5ml" event={"ID":"13c70bc8-3f75-401b-8a96-d3d044e39672","Type":"ContainerDied","Data":"fb6ab8fb89f37fb010e2ca58da098bf09d4aeb72cf5ee3378de079317faf467c"} Mar 11 13:40:13 crc kubenswrapper[4923]: I0311 13:40:13.044986 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553940-wj5ml" Mar 11 13:40:13 crc kubenswrapper[4923]: I0311 13:40:13.203496 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5nj5\" (UniqueName: \"kubernetes.io/projected/13c70bc8-3f75-401b-8a96-d3d044e39672-kube-api-access-t5nj5\") pod \"13c70bc8-3f75-401b-8a96-d3d044e39672\" (UID: \"13c70bc8-3f75-401b-8a96-d3d044e39672\") " Mar 11 13:40:13 crc kubenswrapper[4923]: I0311 13:40:13.216696 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13c70bc8-3f75-401b-8a96-d3d044e39672-kube-api-access-t5nj5" (OuterVolumeSpecName: "kube-api-access-t5nj5") pod "13c70bc8-3f75-401b-8a96-d3d044e39672" (UID: "13c70bc8-3f75-401b-8a96-d3d044e39672"). InnerVolumeSpecName "kube-api-access-t5nj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:40:13 crc kubenswrapper[4923]: I0311 13:40:13.305624 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5nj5\" (UniqueName: \"kubernetes.io/projected/13c70bc8-3f75-401b-8a96-d3d044e39672-kube-api-access-t5nj5\") on node \"crc\" DevicePath \"\"" Mar 11 13:40:13 crc kubenswrapper[4923]: I0311 13:40:13.750973 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553940-wj5ml" event={"ID":"13c70bc8-3f75-401b-8a96-d3d044e39672","Type":"ContainerDied","Data":"5b17fc1e9108d770cef387d2219540ae54822e0433148406be64e168b4ac9f04"} Mar 11 13:40:13 crc kubenswrapper[4923]: I0311 13:40:13.751013 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b17fc1e9108d770cef387d2219540ae54822e0433148406be64e168b4ac9f04" Mar 11 13:40:13 crc kubenswrapper[4923]: I0311 13:40:13.751020 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553940-wj5ml" Mar 11 13:40:14 crc kubenswrapper[4923]: I0311 13:40:14.178077 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29553934-p2kch"] Mar 11 13:40:14 crc kubenswrapper[4923]: I0311 13:40:14.185385 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29553934-p2kch"] Mar 11 13:40:14 crc kubenswrapper[4923]: I0311 13:40:14.763375 4923 generic.go:334] "Generic (PLEG): container finished" podID="f7361c58-e41e-42ab-9c24-d0e470ccea54" containerID="304bb4d25b0ee4a7e44d56d37696f9762d3859438b560ce2b1936ba1a46dd358" exitCode=0 Mar 11 13:40:14 crc kubenswrapper[4923]: I0311 13:40:14.763467 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" event={"ID":"f7361c58-e41e-42ab-9c24-d0e470ccea54","Type":"ContainerDied","Data":"304bb4d25b0ee4a7e44d56d37696f9762d3859438b560ce2b1936ba1a46dd358"} Mar 11 13:40:15 crc kubenswrapper[4923]: I0311 13:40:15.041871 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecd0e139-902f-425f-ba54-08eac773e2ae" path="/var/lib/kubelet/pods/ecd0e139-902f-425f-ba54-08eac773e2ae/volumes" Mar 11 13:40:16 crc kubenswrapper[4923]: I0311 13:40:16.098415 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" Mar 11 13:40:16 crc kubenswrapper[4923]: I0311 13:40:16.252560 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7361c58-e41e-42ab-9c24-d0e470ccea54-scripts\") pod \"f7361c58-e41e-42ab-9c24-d0e470ccea54\" (UID: \"f7361c58-e41e-42ab-9c24-d0e470ccea54\") " Mar 11 13:40:16 crc kubenswrapper[4923]: I0311 13:40:16.252645 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7361c58-e41e-42ab-9c24-d0e470ccea54-config-data\") pod \"f7361c58-e41e-42ab-9c24-d0e470ccea54\" (UID: \"f7361c58-e41e-42ab-9c24-d0e470ccea54\") " Mar 11 13:40:16 crc kubenswrapper[4923]: I0311 13:40:16.252680 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f7361c58-e41e-42ab-9c24-d0e470ccea54-db-sync-config-data\") pod \"f7361c58-e41e-42ab-9c24-d0e470ccea54\" (UID: \"f7361c58-e41e-42ab-9c24-d0e470ccea54\") " Mar 11 13:40:16 crc kubenswrapper[4923]: I0311 13:40:16.252767 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f7361c58-e41e-42ab-9c24-d0e470ccea54-etc-machine-id\") pod \"f7361c58-e41e-42ab-9c24-d0e470ccea54\" (UID: \"f7361c58-e41e-42ab-9c24-d0e470ccea54\") " Mar 11 13:40:16 crc kubenswrapper[4923]: I0311 13:40:16.252859 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xg9h7\" (UniqueName: \"kubernetes.io/projected/f7361c58-e41e-42ab-9c24-d0e470ccea54-kube-api-access-xg9h7\") pod \"f7361c58-e41e-42ab-9c24-d0e470ccea54\" (UID: \"f7361c58-e41e-42ab-9c24-d0e470ccea54\") " Mar 11 13:40:16 crc kubenswrapper[4923]: I0311 13:40:16.253001 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7361c58-e41e-42ab-9c24-d0e470ccea54-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f7361c58-e41e-42ab-9c24-d0e470ccea54" (UID: "f7361c58-e41e-42ab-9c24-d0e470ccea54"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:40:16 crc kubenswrapper[4923]: I0311 13:40:16.253387 4923 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f7361c58-e41e-42ab-9c24-d0e470ccea54-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 11 13:40:16 crc kubenswrapper[4923]: I0311 13:40:16.257837 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7361c58-e41e-42ab-9c24-d0e470ccea54-scripts" (OuterVolumeSpecName: "scripts") pod "f7361c58-e41e-42ab-9c24-d0e470ccea54" (UID: "f7361c58-e41e-42ab-9c24-d0e470ccea54"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:40:16 crc kubenswrapper[4923]: I0311 13:40:16.258029 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7361c58-e41e-42ab-9c24-d0e470ccea54-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f7361c58-e41e-42ab-9c24-d0e470ccea54" (UID: "f7361c58-e41e-42ab-9c24-d0e470ccea54"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:40:16 crc kubenswrapper[4923]: I0311 13:40:16.258842 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7361c58-e41e-42ab-9c24-d0e470ccea54-kube-api-access-xg9h7" (OuterVolumeSpecName: "kube-api-access-xg9h7") pod "f7361c58-e41e-42ab-9c24-d0e470ccea54" (UID: "f7361c58-e41e-42ab-9c24-d0e470ccea54"). InnerVolumeSpecName "kube-api-access-xg9h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:40:16 crc kubenswrapper[4923]: I0311 13:40:16.298186 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7361c58-e41e-42ab-9c24-d0e470ccea54-config-data" (OuterVolumeSpecName: "config-data") pod "f7361c58-e41e-42ab-9c24-d0e470ccea54" (UID: "f7361c58-e41e-42ab-9c24-d0e470ccea54"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:40:16 crc kubenswrapper[4923]: I0311 13:40:16.354152 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f7361c58-e41e-42ab-9c24-d0e470ccea54-scripts\") on node \"crc\" DevicePath \"\"" Mar 11 13:40:16 crc kubenswrapper[4923]: I0311 13:40:16.354180 4923 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f7361c58-e41e-42ab-9c24-d0e470ccea54-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 11 13:40:16 crc kubenswrapper[4923]: I0311 13:40:16.354191 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f7361c58-e41e-42ab-9c24-d0e470ccea54-config-data\") on node \"crc\" DevicePath \"\"" Mar 11 13:40:16 crc kubenswrapper[4923]: I0311 13:40:16.354200 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xg9h7\" (UniqueName: \"kubernetes.io/projected/f7361c58-e41e-42ab-9c24-d0e470ccea54-kube-api-access-xg9h7\") on node \"crc\" DevicePath \"\"" Mar 11 13:40:16 crc kubenswrapper[4923]: I0311 13:40:16.784991 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" event={"ID":"f7361c58-e41e-42ab-9c24-d0e470ccea54","Type":"ContainerDied","Data":"e8c0a3b8167e29c05e43f023027501f07991fa3eefe0315779b1f69101dbdb72"} Mar 11 13:40:16 crc kubenswrapper[4923]: I0311 13:40:16.785050 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-gbxj9" Mar 11 13:40:16 crc kubenswrapper[4923]: I0311 13:40:16.785051 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e8c0a3b8167e29c05e43f023027501f07991fa3eefe0315779b1f69101dbdb72" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.074092 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Mar 11 13:40:17 crc kubenswrapper[4923]: E0311 13:40:17.074713 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13c70bc8-3f75-401b-8a96-d3d044e39672" containerName="oc" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.074738 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="13c70bc8-3f75-401b-8a96-d3d044e39672" containerName="oc" Mar 11 13:40:17 crc kubenswrapper[4923]: E0311 13:40:17.074763 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7361c58-e41e-42ab-9c24-d0e470ccea54" containerName="cinder-db-sync" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.074773 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7361c58-e41e-42ab-9c24-d0e470ccea54" containerName="cinder-db-sync" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.074929 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="13c70bc8-3f75-401b-8a96-d3d044e39672" containerName="oc" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.074943 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7361c58-e41e-42ab-9c24-d0e470ccea54" containerName="cinder-db-sync" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.075920 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.078982 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scheduler-config-data" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.079075 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scripts" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.079200 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-cinder-dockercfg-pkktf" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.093363 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-config-data" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.095569 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.127781 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.128884 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.130976 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-volume-volume1-config-data" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.163496 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.166151 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fce6ff82-c436-41e1-bcef-f16413e39c71-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fce6ff82-c436-41e1-bcef-f16413e39c71\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.166259 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgp9m\" (UniqueName: \"kubernetes.io/projected/fce6ff82-c436-41e1-bcef-f16413e39c71-kube-api-access-wgp9m\") pod \"cinder-scheduler-0\" (UID: \"fce6ff82-c436-41e1-bcef-f16413e39c71\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.166326 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fce6ff82-c436-41e1-bcef-f16413e39c71-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fce6ff82-c436-41e1-bcef-f16413e39c71\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.166364 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fce6ff82-c436-41e1-bcef-f16413e39c71-scripts\") pod \"cinder-scheduler-0\" (UID: \"fce6ff82-c436-41e1-bcef-f16413e39c71\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.166389 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fce6ff82-c436-41e1-bcef-f16413e39c71-config-data\") pod \"cinder-scheduler-0\" (UID: \"fce6ff82-c436-41e1-bcef-f16413e39c71\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.168938 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.175963 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.186533 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-backup-config-data" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.198902 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.246609 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.247754 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.259018 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.260312 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-api-config-data" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.267173 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.267208 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-lib-modules\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.267226 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.267242 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.267429 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.267480 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-run\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.267503 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5qc8\" (UniqueName: \"kubernetes.io/projected/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-kube-api-access-w5qc8\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.267531 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-config-data-custom\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.267554 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.267576 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7hfj\" (UniqueName: \"kubernetes.io/projected/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-kube-api-access-h7hfj\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.267632 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.267748 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-dev\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.267813 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-run\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.267843 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.267898 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.267989 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fce6ff82-c436-41e1-bcef-f16413e39c71-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fce6ff82-c436-41e1-bcef-f16413e39c71\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.268059 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-sys\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.268086 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.268129 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-etc-nvme\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.268159 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgp9m\" (UniqueName: \"kubernetes.io/projected/fce6ff82-c436-41e1-bcef-f16413e39c71-kube-api-access-wgp9m\") pod \"cinder-scheduler-0\" (UID: \"fce6ff82-c436-41e1-bcef-f16413e39c71\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.268181 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-dev\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.268207 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-sys\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.268465 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fce6ff82-c436-41e1-bcef-f16413e39c71-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fce6ff82-c436-41e1-bcef-f16413e39c71\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.268519 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-config-data\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.268534 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fce6ff82-c436-41e1-bcef-f16413e39c71-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fce6ff82-c436-41e1-bcef-f16413e39c71\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.268541 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fce6ff82-c436-41e1-bcef-f16413e39c71-scripts\") pod \"cinder-scheduler-0\" (UID: \"fce6ff82-c436-41e1-bcef-f16413e39c71\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.268591 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-scripts\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.268614 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.268636 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.268658 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fce6ff82-c436-41e1-bcef-f16413e39c71-config-data\") pod \"cinder-scheduler-0\" (UID: \"fce6ff82-c436-41e1-bcef-f16413e39c71\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.268676 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.268694 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.268726 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.268760 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.275633 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fce6ff82-c436-41e1-bcef-f16413e39c71-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fce6ff82-c436-41e1-bcef-f16413e39c71\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.281633 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fce6ff82-c436-41e1-bcef-f16413e39c71-scripts\") pod \"cinder-scheduler-0\" (UID: \"fce6ff82-c436-41e1-bcef-f16413e39c71\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.283222 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fce6ff82-c436-41e1-bcef-f16413e39c71-config-data\") pod \"cinder-scheduler-0\" (UID: \"fce6ff82-c436-41e1-bcef-f16413e39c71\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.308065 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgp9m\" (UniqueName: \"kubernetes.io/projected/fce6ff82-c436-41e1-bcef-f16413e39c71-kube-api-access-wgp9m\") pod \"cinder-scheduler-0\" (UID: \"fce6ff82-c436-41e1-bcef-f16413e39c71\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369489 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369531 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369568 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-sys\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369587 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369608 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrn27\" (UniqueName: \"kubernetes.io/projected/811524db-4494-42cb-a404-25b40a5614f4-kube-api-access-mrn27\") pod \"cinder-api-0\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369604 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369627 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-etc-nvme\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369647 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-dev\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369663 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-sys\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369688 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/811524db-4494-42cb-a404-25b40a5614f4-config-data\") pod \"cinder-api-0\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369717 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-config-data\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369735 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-scripts\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369753 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369774 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369792 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369807 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369823 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369843 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369862 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/811524db-4494-42cb-a404-25b40a5614f4-config-data-custom\") pod \"cinder-api-0\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369885 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/811524db-4494-42cb-a404-25b40a5614f4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369893 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369903 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369927 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370068 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-etc-nvme\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369946 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-dev\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370012 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-sys\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370092 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370035 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370042 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370047 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370048 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.369930 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-sys\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370017 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370387 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370417 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370439 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-lib-modules\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370465 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/811524db-4494-42cb-a404-25b40a5614f4-logs\") pod \"cinder-api-0\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370550 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370585 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-run\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370628 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5qc8\" (UniqueName: \"kubernetes.io/projected/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-kube-api-access-w5qc8\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370654 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-config-data-custom\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370683 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370704 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370724 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7hfj\" (UniqueName: \"kubernetes.io/projected/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-kube-api-access-h7hfj\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370760 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/811524db-4494-42cb-a404-25b40a5614f4-scripts\") pod \"cinder-api-0\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370767 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-run\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370793 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-dev\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370808 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370821 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-run\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370832 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-lib-modules\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370966 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-dev\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.370998 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.371200 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-run\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.371254 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.374234 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.379790 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.380717 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-config-data\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.381575 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-scripts\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.388257 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-config-data-custom\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.388729 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7hfj\" (UniqueName: \"kubernetes.io/projected/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-kube-api-access-h7hfj\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.388933 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.389121 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5qc8\" (UniqueName: \"kubernetes.io/projected/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-kube-api-access-w5qc8\") pod \"cinder-backup-0\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.391373 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.444738 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.472180 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/811524db-4494-42cb-a404-25b40a5614f4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.472405 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/811524db-4494-42cb-a404-25b40a5614f4-logs\") pod \"cinder-api-0\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.472550 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/811524db-4494-42cb-a404-25b40a5614f4-scripts\") pod \"cinder-api-0\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.472687 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrn27\" (UniqueName: \"kubernetes.io/projected/811524db-4494-42cb-a404-25b40a5614f4-kube-api-access-mrn27\") pod \"cinder-api-0\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.472863 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/811524db-4494-42cb-a404-25b40a5614f4-config-data\") pod \"cinder-api-0\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.473001 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/811524db-4494-42cb-a404-25b40a5614f4-config-data-custom\") pod \"cinder-api-0\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.473159 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/811524db-4494-42cb-a404-25b40a5614f4-logs\") pod \"cinder-api-0\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.472571 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/811524db-4494-42cb-a404-25b40a5614f4-etc-machine-id\") pod \"cinder-api-0\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.477241 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/811524db-4494-42cb-a404-25b40a5614f4-scripts\") pod \"cinder-api-0\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.477953 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/811524db-4494-42cb-a404-25b40a5614f4-config-data\") pod \"cinder-api-0\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.483898 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.491191 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrn27\" (UniqueName: \"kubernetes.io/projected/811524db-4494-42cb-a404-25b40a5614f4-kube-api-access-mrn27\") pod \"cinder-api-0\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.492013 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/811524db-4494-42cb-a404-25b40a5614f4-config-data-custom\") pod \"cinder-api-0\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.562577 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.636610 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.793782 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"fce6ff82-c436-41e1-bcef-f16413e39c71","Type":"ContainerStarted","Data":"dd5b8d7960504e138033dd1e12938f82ee1bb160331c04dacf527003aaf53cb9"} Mar 11 13:40:17 crc kubenswrapper[4923]: W0311 13:40:17.894313 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded2ffa6a_46e2_4257_a8c0_54d2808b37c4.slice/crio-7dc3378ac936b4ff8d85d801f595353e896cb0d7d88d141b3d07f19be0ae7571 WatchSource:0}: Error finding container 7dc3378ac936b4ff8d85d801f595353e896cb0d7d88d141b3d07f19be0ae7571: Status 404 returned error can't find the container with id 7dc3378ac936b4ff8d85d801f595353e896cb0d7d88d141b3d07f19be0ae7571 Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.897723 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Mar 11 13:40:17 crc kubenswrapper[4923]: I0311 13:40:17.996490 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Mar 11 13:40:18 crc kubenswrapper[4923]: W0311 13:40:18.000736 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9b2b9ec_01a3_41a7_bf78_143b4be40a0c.slice/crio-30d12174232cc0bd0e2030e72644f8a2297fb7e9c2114fd6281b4caad66e07b9 WatchSource:0}: Error finding container 30d12174232cc0bd0e2030e72644f8a2297fb7e9c2114fd6281b4caad66e07b9: Status 404 returned error can't find the container with id 30d12174232cc0bd0e2030e72644f8a2297fb7e9c2114fd6281b4caad66e07b9 Mar 11 13:40:18 crc kubenswrapper[4923]: I0311 13:40:18.050932 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Mar 11 13:40:18 crc kubenswrapper[4923]: W0311 13:40:18.061428 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod811524db_4494_42cb_a404_25b40a5614f4.slice/crio-e56961fcfe8fc2210d49892cf2e92d85fbabc1b4ca06b36d9d1058b305dcda40 WatchSource:0}: Error finding container e56961fcfe8fc2210d49892cf2e92d85fbabc1b4ca06b36d9d1058b305dcda40: Status 404 returned error can't find the container with id e56961fcfe8fc2210d49892cf2e92d85fbabc1b4ca06b36d9d1058b305dcda40 Mar 11 13:40:18 crc kubenswrapper[4923]: I0311 13:40:18.803106 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"811524db-4494-42cb-a404-25b40a5614f4","Type":"ContainerStarted","Data":"a3bad500be86c9560cb179786f29d771d52a094d4367d0242c26c4822fdd0626"} Mar 11 13:40:18 crc kubenswrapper[4923]: I0311 13:40:18.803454 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"811524db-4494-42cb-a404-25b40a5614f4","Type":"ContainerStarted","Data":"e56961fcfe8fc2210d49892cf2e92d85fbabc1b4ca06b36d9d1058b305dcda40"} Mar 11 13:40:18 crc kubenswrapper[4923]: I0311 13:40:18.804278 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c","Type":"ContainerStarted","Data":"30d12174232cc0bd0e2030e72644f8a2297fb7e9c2114fd6281b4caad66e07b9"} Mar 11 13:40:18 crc kubenswrapper[4923]: I0311 13:40:18.805124 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4","Type":"ContainerStarted","Data":"7dc3378ac936b4ff8d85d801f595353e896cb0d7d88d141b3d07f19be0ae7571"} Mar 11 13:40:19 crc kubenswrapper[4923]: I0311 13:40:19.817794 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c","Type":"ContainerStarted","Data":"882442eca39bb3e6fa387944706810da0348a93c5fd04094e8b1de3cecc8f208"} Mar 11 13:40:19 crc kubenswrapper[4923]: I0311 13:40:19.818459 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c","Type":"ContainerStarted","Data":"07d0c4907846fe6ec0f36a2b68ef7560e6274941e23d6b568ed6d0374c8ebad0"} Mar 11 13:40:19 crc kubenswrapper[4923]: I0311 13:40:19.820188 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4","Type":"ContainerStarted","Data":"557f53ab705a82805bb730cf15949a3489567ca6f7a2c90347da74487ec25cd9"} Mar 11 13:40:19 crc kubenswrapper[4923]: I0311 13:40:19.820301 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4","Type":"ContainerStarted","Data":"ba7f155682187e45a107fb30addc97abda03e4e9551b321ff11260a995f13d38"} Mar 11 13:40:19 crc kubenswrapper[4923]: I0311 13:40:19.823476 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"fce6ff82-c436-41e1-bcef-f16413e39c71","Type":"ContainerStarted","Data":"0457653217cc4be1bc77252e1a96fe9be0aa46fda288ff8562e108c88e27d142"} Mar 11 13:40:19 crc kubenswrapper[4923]: I0311 13:40:19.823514 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"fce6ff82-c436-41e1-bcef-f16413e39c71","Type":"ContainerStarted","Data":"196c02cd4bb53f4ecfc1a4200e37811d9001d1730c33f790c0922488994bf296"} Mar 11 13:40:19 crc kubenswrapper[4923]: I0311 13:40:19.825530 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"811524db-4494-42cb-a404-25b40a5614f4","Type":"ContainerStarted","Data":"34ad1d334751d655aa2bd86e938ad9100f02f4969d25c60fd94512e6c86896ed"} Mar 11 13:40:19 crc kubenswrapper[4923]: I0311 13:40:19.826292 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:40:19 crc kubenswrapper[4923]: I0311 13:40:19.844573 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-backup-0" podStartSLOduration=2.042321626 podStartE2EDuration="2.844558241s" podCreationTimestamp="2026-03-11 13:40:17 +0000 UTC" firstStartedPulling="2026-03-11 13:40:18.003779604 +0000 UTC m=+1065.685681318" lastFinishedPulling="2026-03-11 13:40:18.806016219 +0000 UTC m=+1066.487917933" observedRunningTime="2026-03-11 13:40:19.838794495 +0000 UTC m=+1067.520696209" watchObservedRunningTime="2026-03-11 13:40:19.844558241 +0000 UTC m=+1067.526459955" Mar 11 13:40:19 crc kubenswrapper[4923]: I0311 13:40:19.859811 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-scheduler-0" podStartSLOduration=2.055544451 podStartE2EDuration="2.859791937s" podCreationTimestamp="2026-03-11 13:40:17 +0000 UTC" firstStartedPulling="2026-03-11 13:40:17.653145401 +0000 UTC m=+1065.335047115" lastFinishedPulling="2026-03-11 13:40:18.457392887 +0000 UTC m=+1066.139294601" observedRunningTime="2026-03-11 13:40:19.856485133 +0000 UTC m=+1067.538386857" watchObservedRunningTime="2026-03-11 13:40:19.859791937 +0000 UTC m=+1067.541693651" Mar 11 13:40:19 crc kubenswrapper[4923]: I0311 13:40:19.884304 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podStartSLOduration=1.643294007 podStartE2EDuration="2.884288358s" podCreationTimestamp="2026-03-11 13:40:17 +0000 UTC" firstStartedPulling="2026-03-11 13:40:17.896559138 +0000 UTC m=+1065.578460872" lastFinishedPulling="2026-03-11 13:40:19.137553509 +0000 UTC m=+1066.819455223" observedRunningTime="2026-03-11 13:40:19.882103652 +0000 UTC m=+1067.564005376" watchObservedRunningTime="2026-03-11 13:40:19.884288358 +0000 UTC m=+1067.566190072" Mar 11 13:40:21 crc kubenswrapper[4923]: I0311 13:40:21.858699 4923 generic.go:334] "Generic (PLEG): container finished" podID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerID="557f53ab705a82805bb730cf15949a3489567ca6f7a2c90347da74487ec25cd9" exitCode=1 Mar 11 13:40:21 crc kubenswrapper[4923]: I0311 13:40:21.858796 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4","Type":"ContainerDied","Data":"557f53ab705a82805bb730cf15949a3489567ca6f7a2c90347da74487ec25cd9"} Mar 11 13:40:21 crc kubenswrapper[4923]: I0311 13:40:21.859966 4923 scope.go:117] "RemoveContainer" containerID="557f53ab705a82805bb730cf15949a3489567ca6f7a2c90347da74487ec25cd9" Mar 11 13:40:21 crc kubenswrapper[4923]: I0311 13:40:21.892331 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-0" podStartSLOduration=4.892308253 podStartE2EDuration="4.892308253s" podCreationTimestamp="2026-03-11 13:40:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:40:19.906719446 +0000 UTC m=+1067.588621150" watchObservedRunningTime="2026-03-11 13:40:21.892308253 +0000 UTC m=+1069.574209997" Mar 11 13:40:22 crc kubenswrapper[4923]: I0311 13:40:22.870522 4923 generic.go:334] "Generic (PLEG): container finished" podID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerID="ba7f155682187e45a107fb30addc97abda03e4e9551b321ff11260a995f13d38" exitCode=1 Mar 11 13:40:22 crc kubenswrapper[4923]: I0311 13:40:22.870618 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4","Type":"ContainerDied","Data":"ba7f155682187e45a107fb30addc97abda03e4e9551b321ff11260a995f13d38"} Mar 11 13:40:24 crc kubenswrapper[4923]: I0311 13:40:24.907010 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4","Type":"ContainerStarted","Data":"495d15e5c1aaefacef2747407bae37ae086143f8f7ad6cf53681dadb263daa0a"} Mar 11 13:40:24 crc kubenswrapper[4923]: I0311 13:40:24.907956 4923 scope.go:117] "RemoveContainer" containerID="ba7f155682187e45a107fb30addc97abda03e4e9551b321ff11260a995f13d38" Mar 11 13:40:25 crc kubenswrapper[4923]: I0311 13:40:25.919231 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4","Type":"ContainerStarted","Data":"9b6f74d17779d01a178bbc820dce465fadd5f07ee130b3bdbf4ee939c0ae4966"} Mar 11 13:40:26 crc kubenswrapper[4923]: I0311 13:40:26.929595 4923 generic.go:334] "Generic (PLEG): container finished" podID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerID="9b6f74d17779d01a178bbc820dce465fadd5f07ee130b3bdbf4ee939c0ae4966" exitCode=1 Mar 11 13:40:26 crc kubenswrapper[4923]: I0311 13:40:26.929627 4923 generic.go:334] "Generic (PLEG): container finished" podID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerID="495d15e5c1aaefacef2747407bae37ae086143f8f7ad6cf53681dadb263daa0a" exitCode=1 Mar 11 13:40:26 crc kubenswrapper[4923]: I0311 13:40:26.929646 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4","Type":"ContainerDied","Data":"9b6f74d17779d01a178bbc820dce465fadd5f07ee130b3bdbf4ee939c0ae4966"} Mar 11 13:40:26 crc kubenswrapper[4923]: I0311 13:40:26.929674 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4","Type":"ContainerDied","Data":"495d15e5c1aaefacef2747407bae37ae086143f8f7ad6cf53681dadb263daa0a"} Mar 11 13:40:26 crc kubenswrapper[4923]: I0311 13:40:26.929693 4923 scope.go:117] "RemoveContainer" containerID="ba7f155682187e45a107fb30addc97abda03e4e9551b321ff11260a995f13d38" Mar 11 13:40:26 crc kubenswrapper[4923]: I0311 13:40:26.931022 4923 scope.go:117] "RemoveContainer" containerID="9b6f74d17779d01a178bbc820dce465fadd5f07ee130b3bdbf4ee939c0ae4966" Mar 11 13:40:26 crc kubenswrapper[4923]: I0311 13:40:26.931143 4923 scope.go:117] "RemoveContainer" containerID="495d15e5c1aaefacef2747407bae37ae086143f8f7ad6cf53681dadb263daa0a" Mar 11 13:40:26 crc kubenswrapper[4923]: E0311 13:40:26.932108 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:40:26 crc kubenswrapper[4923]: I0311 13:40:26.980081 4923 scope.go:117] "RemoveContainer" containerID="557f53ab705a82805bb730cf15949a3489567ca6f7a2c90347da74487ec25cd9" Mar 11 13:40:27 crc kubenswrapper[4923]: I0311 13:40:27.392003 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:40:27 crc kubenswrapper[4923]: I0311 13:40:27.445543 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:27 crc kubenswrapper[4923]: I0311 13:40:27.445588 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:27 crc kubenswrapper[4923]: I0311 13:40:27.484530 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:27 crc kubenswrapper[4923]: I0311 13:40:27.940674 4923 scope.go:117] "RemoveContainer" containerID="9b6f74d17779d01a178bbc820dce465fadd5f07ee130b3bdbf4ee939c0ae4966" Mar 11 13:40:27 crc kubenswrapper[4923]: I0311 13:40:27.940719 4923 scope.go:117] "RemoveContainer" containerID="495d15e5c1aaefacef2747407bae37ae086143f8f7ad6cf53681dadb263daa0a" Mar 11 13:40:27 crc kubenswrapper[4923]: E0311 13:40:27.941099 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:40:28 crc kubenswrapper[4923]: I0311 13:40:28.946035 4923 scope.go:117] "RemoveContainer" containerID="9b6f74d17779d01a178bbc820dce465fadd5f07ee130b3bdbf4ee939c0ae4966" Mar 11 13:40:28 crc kubenswrapper[4923]: I0311 13:40:28.946311 4923 scope.go:117] "RemoveContainer" containerID="495d15e5c1aaefacef2747407bae37ae086143f8f7ad6cf53681dadb263daa0a" Mar 11 13:40:28 crc kubenswrapper[4923]: E0311 13:40:28.946552 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:40:29 crc kubenswrapper[4923]: I0311 13:40:29.445800 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:29 crc kubenswrapper[4923]: I0311 13:40:29.953631 4923 scope.go:117] "RemoveContainer" containerID="9b6f74d17779d01a178bbc820dce465fadd5f07ee130b3bdbf4ee939c0ae4966" Mar 11 13:40:29 crc kubenswrapper[4923]: I0311 13:40:29.953685 4923 scope.go:117] "RemoveContainer" containerID="495d15e5c1aaefacef2747407bae37ae086143f8f7ad6cf53681dadb263daa0a" Mar 11 13:40:29 crc kubenswrapper[4923]: E0311 13:40:29.954147 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 10s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 10s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:40:33 crc kubenswrapper[4923]: I0311 13:40:33.898974 4923 scope.go:117] "RemoveContainer" containerID="922fad3b003dde7f78c678afe06de492a98a65c1a6db57c64cf032ab8918dc52" Mar 11 13:40:37 crc kubenswrapper[4923]: I0311 13:40:37.568971 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:40:37 crc kubenswrapper[4923]: I0311 13:40:37.724169 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:40:39 crc kubenswrapper[4923]: I0311 13:40:39.416998 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:40:41 crc kubenswrapper[4923]: I0311 13:40:41.932604 4923 patch_prober.go:28] interesting pod/machine-config-daemon-fsfnw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 11 13:40:41 crc kubenswrapper[4923]: I0311 13:40:41.933580 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 11 13:40:43 crc kubenswrapper[4923]: I0311 13:40:43.040589 4923 scope.go:117] "RemoveContainer" containerID="9b6f74d17779d01a178bbc820dce465fadd5f07ee130b3bdbf4ee939c0ae4966" Mar 11 13:40:43 crc kubenswrapper[4923]: I0311 13:40:43.041932 4923 scope.go:117] "RemoveContainer" containerID="495d15e5c1aaefacef2747407bae37ae086143f8f7ad6cf53681dadb263daa0a" Mar 11 13:40:44 crc kubenswrapper[4923]: I0311 13:40:44.072805 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4","Type":"ContainerStarted","Data":"3f40bd40ac03938fbaff964efdcb6049615cf519cdf8742dff753938d0e2a5c8"} Mar 11 13:40:44 crc kubenswrapper[4923]: I0311 13:40:44.073368 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4","Type":"ContainerStarted","Data":"3be97aaa512e754be33cb25034d5324bffb339b2825fc3ecc25af4b028e7c57f"} Mar 11 13:40:45 crc kubenswrapper[4923]: I0311 13:40:45.090508 4923 generic.go:334] "Generic (PLEG): container finished" podID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerID="3f40bd40ac03938fbaff964efdcb6049615cf519cdf8742dff753938d0e2a5c8" exitCode=1 Mar 11 13:40:45 crc kubenswrapper[4923]: I0311 13:40:45.090594 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4","Type":"ContainerDied","Data":"3f40bd40ac03938fbaff964efdcb6049615cf519cdf8742dff753938d0e2a5c8"} Mar 11 13:40:45 crc kubenswrapper[4923]: I0311 13:40:45.090764 4923 scope.go:117] "RemoveContainer" containerID="495d15e5c1aaefacef2747407bae37ae086143f8f7ad6cf53681dadb263daa0a" Mar 11 13:40:45 crc kubenswrapper[4923]: I0311 13:40:45.091299 4923 scope.go:117] "RemoveContainer" containerID="3f40bd40ac03938fbaff964efdcb6049615cf519cdf8742dff753938d0e2a5c8" Mar 11 13:40:45 crc kubenswrapper[4923]: E0311 13:40:45.091550 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:40:47 crc kubenswrapper[4923]: I0311 13:40:47.111407 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4","Type":"ContainerDied","Data":"3be97aaa512e754be33cb25034d5324bffb339b2825fc3ecc25af4b028e7c57f"} Mar 11 13:40:47 crc kubenswrapper[4923]: I0311 13:40:47.111748 4923 scope.go:117] "RemoveContainer" containerID="9b6f74d17779d01a178bbc820dce465fadd5f07ee130b3bdbf4ee939c0ae4966" Mar 11 13:40:47 crc kubenswrapper[4923]: I0311 13:40:47.111332 4923 generic.go:334] "Generic (PLEG): container finished" podID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerID="3be97aaa512e754be33cb25034d5324bffb339b2825fc3ecc25af4b028e7c57f" exitCode=1 Mar 11 13:40:47 crc kubenswrapper[4923]: I0311 13:40:47.112307 4923 scope.go:117] "RemoveContainer" containerID="3be97aaa512e754be33cb25034d5324bffb339b2825fc3ecc25af4b028e7c57f" Mar 11 13:40:47 crc kubenswrapper[4923]: I0311 13:40:47.112389 4923 scope.go:117] "RemoveContainer" containerID="3f40bd40ac03938fbaff964efdcb6049615cf519cdf8742dff753938d0e2a5c8" Mar 11 13:40:47 crc kubenswrapper[4923]: E0311 13:40:47.112682 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:40:47 crc kubenswrapper[4923]: I0311 13:40:47.445461 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:47 crc kubenswrapper[4923]: I0311 13:40:47.445553 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:47 crc kubenswrapper[4923]: I0311 13:40:47.445577 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:40:48 crc kubenswrapper[4923]: I0311 13:40:48.121098 4923 scope.go:117] "RemoveContainer" containerID="3be97aaa512e754be33cb25034d5324bffb339b2825fc3ecc25af4b028e7c57f" Mar 11 13:40:48 crc kubenswrapper[4923]: I0311 13:40:48.121756 4923 scope.go:117] "RemoveContainer" containerID="3f40bd40ac03938fbaff964efdcb6049615cf519cdf8742dff753938d0e2a5c8" Mar 11 13:40:48 crc kubenswrapper[4923]: E0311 13:40:48.122110 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:41:01 crc kubenswrapper[4923]: I0311 13:41:01.022567 4923 scope.go:117] "RemoveContainer" containerID="3be97aaa512e754be33cb25034d5324bffb339b2825fc3ecc25af4b028e7c57f" Mar 11 13:41:01 crc kubenswrapper[4923]: I0311 13:41:01.023146 4923 scope.go:117] "RemoveContainer" containerID="3f40bd40ac03938fbaff964efdcb6049615cf519cdf8742dff753938d0e2a5c8" Mar 11 13:41:01 crc kubenswrapper[4923]: E0311 13:41:01.023456 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:41:11 crc kubenswrapper[4923]: I0311 13:41:11.932691 4923 patch_prober.go:28] interesting pod/machine-config-daemon-fsfnw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 11 13:41:11 crc kubenswrapper[4923]: I0311 13:41:11.933282 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 11 13:41:12 crc kubenswrapper[4923]: I0311 13:41:12.023058 4923 scope.go:117] "RemoveContainer" containerID="3be97aaa512e754be33cb25034d5324bffb339b2825fc3ecc25af4b028e7c57f" Mar 11 13:41:12 crc kubenswrapper[4923]: I0311 13:41:12.023099 4923 scope.go:117] "RemoveContainer" containerID="3f40bd40ac03938fbaff964efdcb6049615cf519cdf8742dff753938d0e2a5c8" Mar 11 13:41:12 crc kubenswrapper[4923]: I0311 13:41:12.677771 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4","Type":"ContainerStarted","Data":"76bc1f6f3d082a7db141d48526186c3eb283fefdfbb9d95d71884ae12ab7e407"} Mar 11 13:41:12 crc kubenswrapper[4923]: I0311 13:41:12.678488 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4","Type":"ContainerStarted","Data":"43efca9fc88fa5e0648576fa09376a3ac49863b60254c0de0dfbc1ee6dee21f9"} Mar 11 13:41:14 crc kubenswrapper[4923]: I0311 13:41:14.698576 4923 generic.go:334] "Generic (PLEG): container finished" podID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerID="76bc1f6f3d082a7db141d48526186c3eb283fefdfbb9d95d71884ae12ab7e407" exitCode=1 Mar 11 13:41:14 crc kubenswrapper[4923]: I0311 13:41:14.698680 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4","Type":"ContainerDied","Data":"76bc1f6f3d082a7db141d48526186c3eb283fefdfbb9d95d71884ae12ab7e407"} Mar 11 13:41:14 crc kubenswrapper[4923]: I0311 13:41:14.699248 4923 scope.go:117] "RemoveContainer" containerID="3f40bd40ac03938fbaff964efdcb6049615cf519cdf8742dff753938d0e2a5c8" Mar 11 13:41:14 crc kubenswrapper[4923]: I0311 13:41:14.700146 4923 scope.go:117] "RemoveContainer" containerID="76bc1f6f3d082a7db141d48526186c3eb283fefdfbb9d95d71884ae12ab7e407" Mar 11 13:41:14 crc kubenswrapper[4923]: E0311 13:41:14.700540 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:41:15 crc kubenswrapper[4923]: I0311 13:41:15.712098 4923 generic.go:334] "Generic (PLEG): container finished" podID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerID="43efca9fc88fa5e0648576fa09376a3ac49863b60254c0de0dfbc1ee6dee21f9" exitCode=1 Mar 11 13:41:15 crc kubenswrapper[4923]: I0311 13:41:15.712146 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4","Type":"ContainerDied","Data":"43efca9fc88fa5e0648576fa09376a3ac49863b60254c0de0dfbc1ee6dee21f9"} Mar 11 13:41:15 crc kubenswrapper[4923]: I0311 13:41:15.712184 4923 scope.go:117] "RemoveContainer" containerID="3be97aaa512e754be33cb25034d5324bffb339b2825fc3ecc25af4b028e7c57f" Mar 11 13:41:15 crc kubenswrapper[4923]: I0311 13:41:15.713797 4923 scope.go:117] "RemoveContainer" containerID="43efca9fc88fa5e0648576fa09376a3ac49863b60254c0de0dfbc1ee6dee21f9" Mar 11 13:41:15 crc kubenswrapper[4923]: I0311 13:41:15.713861 4923 scope.go:117] "RemoveContainer" containerID="76bc1f6f3d082a7db141d48526186c3eb283fefdfbb9d95d71884ae12ab7e407" Mar 11 13:41:15 crc kubenswrapper[4923]: E0311 13:41:15.714535 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:41:17 crc kubenswrapper[4923]: I0311 13:41:17.445378 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:41:17 crc kubenswrapper[4923]: I0311 13:41:17.445713 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:41:17 crc kubenswrapper[4923]: I0311 13:41:17.445735 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:41:17 crc kubenswrapper[4923]: I0311 13:41:17.446686 4923 scope.go:117] "RemoveContainer" containerID="43efca9fc88fa5e0648576fa09376a3ac49863b60254c0de0dfbc1ee6dee21f9" Mar 11 13:41:17 crc kubenswrapper[4923]: I0311 13:41:17.446715 4923 scope.go:117] "RemoveContainer" containerID="76bc1f6f3d082a7db141d48526186c3eb283fefdfbb9d95d71884ae12ab7e407" Mar 11 13:41:17 crc kubenswrapper[4923]: E0311 13:41:17.447108 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:41:31 crc kubenswrapper[4923]: I0311 13:41:31.023081 4923 scope.go:117] "RemoveContainer" containerID="43efca9fc88fa5e0648576fa09376a3ac49863b60254c0de0dfbc1ee6dee21f9" Mar 11 13:41:31 crc kubenswrapper[4923]: I0311 13:41:31.023718 4923 scope.go:117] "RemoveContainer" containerID="76bc1f6f3d082a7db141d48526186c3eb283fefdfbb9d95d71884ae12ab7e407" Mar 11 13:41:31 crc kubenswrapper[4923]: E0311 13:41:31.024196 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:41:41 crc kubenswrapper[4923]: I0311 13:41:41.932794 4923 patch_prober.go:28] interesting pod/machine-config-daemon-fsfnw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 11 13:41:41 crc kubenswrapper[4923]: I0311 13:41:41.933496 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 11 13:41:41 crc kubenswrapper[4923]: I0311 13:41:41.933583 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" Mar 11 13:41:41 crc kubenswrapper[4923]: I0311 13:41:41.934546 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ead0c2f701ab9961c880f0090716c64783b2bee2562b04e2163c46d120ab88f4"} pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 11 13:41:41 crc kubenswrapper[4923]: I0311 13:41:41.934644 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" containerID="cri-o://ead0c2f701ab9961c880f0090716c64783b2bee2562b04e2163c46d120ab88f4" gracePeriod=600 Mar 11 13:41:42 crc kubenswrapper[4923]: I0311 13:41:42.022411 4923 scope.go:117] "RemoveContainer" containerID="43efca9fc88fa5e0648576fa09376a3ac49863b60254c0de0dfbc1ee6dee21f9" Mar 11 13:41:42 crc kubenswrapper[4923]: I0311 13:41:42.022437 4923 scope.go:117] "RemoveContainer" containerID="76bc1f6f3d082a7db141d48526186c3eb283fefdfbb9d95d71884ae12ab7e407" Mar 11 13:41:42 crc kubenswrapper[4923]: E0311 13:41:42.022712 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:41:42 crc kubenswrapper[4923]: I0311 13:41:42.980651 4923 generic.go:334] "Generic (PLEG): container finished" podID="356de301-ded3-4112-8406-1a28e44c2082" containerID="ead0c2f701ab9961c880f0090716c64783b2bee2562b04e2163c46d120ab88f4" exitCode=0 Mar 11 13:41:42 crc kubenswrapper[4923]: I0311 13:41:42.980775 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" event={"ID":"356de301-ded3-4112-8406-1a28e44c2082","Type":"ContainerDied","Data":"ead0c2f701ab9961c880f0090716c64783b2bee2562b04e2163c46d120ab88f4"} Mar 11 13:41:42 crc kubenswrapper[4923]: I0311 13:41:42.981021 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" event={"ID":"356de301-ded3-4112-8406-1a28e44c2082","Type":"ContainerStarted","Data":"3d359e768a3b56caad3d49994ffb4f7bcb69d87648b877548ba08cfe716e455c"} Mar 11 13:41:42 crc kubenswrapper[4923]: I0311 13:41:42.981052 4923 scope.go:117] "RemoveContainer" containerID="a555ce4420a87a2fc08d149d6004aea3fd8a76f007c748edd82c76ca49ddb60a" Mar 11 13:41:53 crc kubenswrapper[4923]: I0311 13:41:53.031252 4923 scope.go:117] "RemoveContainer" containerID="43efca9fc88fa5e0648576fa09376a3ac49863b60254c0de0dfbc1ee6dee21f9" Mar 11 13:41:53 crc kubenswrapper[4923]: I0311 13:41:53.031862 4923 scope.go:117] "RemoveContainer" containerID="76bc1f6f3d082a7db141d48526186c3eb283fefdfbb9d95d71884ae12ab7e407" Mar 11 13:41:53 crc kubenswrapper[4923]: E0311 13:41:53.032240 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 40s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:42:00 crc kubenswrapper[4923]: I0311 13:42:00.158174 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29553942-46xs7"] Mar 11 13:42:00 crc kubenswrapper[4923]: I0311 13:42:00.160033 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553942-46xs7" Mar 11 13:42:00 crc kubenswrapper[4923]: I0311 13:42:00.163865 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553942-46xs7"] Mar 11 13:42:00 crc kubenswrapper[4923]: I0311 13:42:00.207623 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 11 13:42:00 crc kubenswrapper[4923]: I0311 13:42:00.207656 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 11 13:42:00 crc kubenswrapper[4923]: I0311 13:42:00.207712 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-nxck7" Mar 11 13:42:00 crc kubenswrapper[4923]: I0311 13:42:00.307101 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5ld5\" (UniqueName: \"kubernetes.io/projected/06883e63-9dba-48dd-a356-d990b50293ff-kube-api-access-h5ld5\") pod \"auto-csr-approver-29553942-46xs7\" (UID: \"06883e63-9dba-48dd-a356-d990b50293ff\") " pod="openshift-infra/auto-csr-approver-29553942-46xs7" Mar 11 13:42:00 crc kubenswrapper[4923]: I0311 13:42:00.409772 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5ld5\" (UniqueName: \"kubernetes.io/projected/06883e63-9dba-48dd-a356-d990b50293ff-kube-api-access-h5ld5\") pod \"auto-csr-approver-29553942-46xs7\" (UID: \"06883e63-9dba-48dd-a356-d990b50293ff\") " pod="openshift-infra/auto-csr-approver-29553942-46xs7" Mar 11 13:42:00 crc kubenswrapper[4923]: I0311 13:42:00.448417 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5ld5\" (UniqueName: \"kubernetes.io/projected/06883e63-9dba-48dd-a356-d990b50293ff-kube-api-access-h5ld5\") pod \"auto-csr-approver-29553942-46xs7\" (UID: \"06883e63-9dba-48dd-a356-d990b50293ff\") " pod="openshift-infra/auto-csr-approver-29553942-46xs7" Mar 11 13:42:00 crc kubenswrapper[4923]: I0311 13:42:00.526512 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553942-46xs7" Mar 11 13:42:00 crc kubenswrapper[4923]: I0311 13:42:00.962805 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553942-46xs7"] Mar 11 13:42:00 crc kubenswrapper[4923]: W0311 13:42:00.967029 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06883e63_9dba_48dd_a356_d990b50293ff.slice/crio-6c0f0e7f559eefdca9836d6194a7dffcc0335d38341da73d4b81c551623fabff WatchSource:0}: Error finding container 6c0f0e7f559eefdca9836d6194a7dffcc0335d38341da73d4b81c551623fabff: Status 404 returned error can't find the container with id 6c0f0e7f559eefdca9836d6194a7dffcc0335d38341da73d4b81c551623fabff Mar 11 13:42:00 crc kubenswrapper[4923]: I0311 13:42:00.971066 4923 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 11 13:42:01 crc kubenswrapper[4923]: I0311 13:42:01.218867 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553942-46xs7" event={"ID":"06883e63-9dba-48dd-a356-d990b50293ff","Type":"ContainerStarted","Data":"6c0f0e7f559eefdca9836d6194a7dffcc0335d38341da73d4b81c551623fabff"} Mar 11 13:42:03 crc kubenswrapper[4923]: I0311 13:42:03.237849 4923 generic.go:334] "Generic (PLEG): container finished" podID="06883e63-9dba-48dd-a356-d990b50293ff" containerID="b5f0c27328a332000af4dd0ed02cbf39e42d2d9095fd9b820964ede98e48c266" exitCode=0 Mar 11 13:42:03 crc kubenswrapper[4923]: I0311 13:42:03.237913 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553942-46xs7" event={"ID":"06883e63-9dba-48dd-a356-d990b50293ff","Type":"ContainerDied","Data":"b5f0c27328a332000af4dd0ed02cbf39e42d2d9095fd9b820964ede98e48c266"} Mar 11 13:42:04 crc kubenswrapper[4923]: I0311 13:42:04.545569 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553942-46xs7" Mar 11 13:42:04 crc kubenswrapper[4923]: I0311 13:42:04.677801 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5ld5\" (UniqueName: \"kubernetes.io/projected/06883e63-9dba-48dd-a356-d990b50293ff-kube-api-access-h5ld5\") pod \"06883e63-9dba-48dd-a356-d990b50293ff\" (UID: \"06883e63-9dba-48dd-a356-d990b50293ff\") " Mar 11 13:42:04 crc kubenswrapper[4923]: I0311 13:42:04.685426 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06883e63-9dba-48dd-a356-d990b50293ff-kube-api-access-h5ld5" (OuterVolumeSpecName: "kube-api-access-h5ld5") pod "06883e63-9dba-48dd-a356-d990b50293ff" (UID: "06883e63-9dba-48dd-a356-d990b50293ff"). InnerVolumeSpecName "kube-api-access-h5ld5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:42:04 crc kubenswrapper[4923]: I0311 13:42:04.779907 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5ld5\" (UniqueName: \"kubernetes.io/projected/06883e63-9dba-48dd-a356-d990b50293ff-kube-api-access-h5ld5\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:05 crc kubenswrapper[4923]: I0311 13:42:05.255584 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553942-46xs7" event={"ID":"06883e63-9dba-48dd-a356-d990b50293ff","Type":"ContainerDied","Data":"6c0f0e7f559eefdca9836d6194a7dffcc0335d38341da73d4b81c551623fabff"} Mar 11 13:42:05 crc kubenswrapper[4923]: I0311 13:42:05.255673 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c0f0e7f559eefdca9836d6194a7dffcc0335d38341da73d4b81c551623fabff" Mar 11 13:42:05 crc kubenswrapper[4923]: I0311 13:42:05.255746 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553942-46xs7" Mar 11 13:42:05 crc kubenswrapper[4923]: I0311 13:42:05.635916 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29553936-k7s5k"] Mar 11 13:42:05 crc kubenswrapper[4923]: I0311 13:42:05.644046 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29553936-k7s5k"] Mar 11 13:42:06 crc kubenswrapper[4923]: I0311 13:42:06.022145 4923 scope.go:117] "RemoveContainer" containerID="43efca9fc88fa5e0648576fa09376a3ac49863b60254c0de0dfbc1ee6dee21f9" Mar 11 13:42:06 crc kubenswrapper[4923]: I0311 13:42:06.022553 4923 scope.go:117] "RemoveContainer" containerID="76bc1f6f3d082a7db141d48526186c3eb283fefdfbb9d95d71884ae12ab7e407" Mar 11 13:42:06 crc kubenswrapper[4923]: I0311 13:42:06.272867 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4","Type":"ContainerStarted","Data":"8bf10a9ec7b294d63184d7c43bf1a0508fd8b54697b5139a5716aae8a4ba506c"} Mar 11 13:42:07 crc kubenswrapper[4923]: I0311 13:42:07.037533 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85eb3ab9-5384-4c51-b20c-66eaa4f44c38" path="/var/lib/kubelet/pods/85eb3ab9-5384-4c51-b20c-66eaa4f44c38/volumes" Mar 11 13:42:07 crc kubenswrapper[4923]: I0311 13:42:07.284765 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4","Type":"ContainerStarted","Data":"4f323cc06c832423a935bd57211a21d27ae7dfdd630767d0c158db5de4fb1394"} Mar 11 13:42:07 crc kubenswrapper[4923]: I0311 13:42:07.445681 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:42:08 crc kubenswrapper[4923]: I0311 13:42:08.297230 4923 generic.go:334] "Generic (PLEG): container finished" podID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerID="4f323cc06c832423a935bd57211a21d27ae7dfdd630767d0c158db5de4fb1394" exitCode=1 Mar 11 13:42:08 crc kubenswrapper[4923]: I0311 13:42:08.297294 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4","Type":"ContainerDied","Data":"4f323cc06c832423a935bd57211a21d27ae7dfdd630767d0c158db5de4fb1394"} Mar 11 13:42:08 crc kubenswrapper[4923]: I0311 13:42:08.297374 4923 scope.go:117] "RemoveContainer" containerID="76bc1f6f3d082a7db141d48526186c3eb283fefdfbb9d95d71884ae12ab7e407" Mar 11 13:42:08 crc kubenswrapper[4923]: I0311 13:42:08.298171 4923 scope.go:117] "RemoveContainer" containerID="4f323cc06c832423a935bd57211a21d27ae7dfdd630767d0c158db5de4fb1394" Mar 11 13:42:08 crc kubenswrapper[4923]: E0311 13:42:08.298529 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:42:09 crc kubenswrapper[4923]: I0311 13:42:09.323806 4923 generic.go:334] "Generic (PLEG): container finished" podID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerID="8bf10a9ec7b294d63184d7c43bf1a0508fd8b54697b5139a5716aae8a4ba506c" exitCode=1 Mar 11 13:42:09 crc kubenswrapper[4923]: I0311 13:42:09.324081 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4","Type":"ContainerDied","Data":"8bf10a9ec7b294d63184d7c43bf1a0508fd8b54697b5139a5716aae8a4ba506c"} Mar 11 13:42:09 crc kubenswrapper[4923]: I0311 13:42:09.324239 4923 scope.go:117] "RemoveContainer" containerID="43efca9fc88fa5e0648576fa09376a3ac49863b60254c0de0dfbc1ee6dee21f9" Mar 11 13:42:09 crc kubenswrapper[4923]: I0311 13:42:09.324312 4923 scope.go:117] "RemoveContainer" containerID="8bf10a9ec7b294d63184d7c43bf1a0508fd8b54697b5139a5716aae8a4ba506c" Mar 11 13:42:09 crc kubenswrapper[4923]: I0311 13:42:09.324358 4923 scope.go:117] "RemoveContainer" containerID="4f323cc06c832423a935bd57211a21d27ae7dfdd630767d0c158db5de4fb1394" Mar 11 13:42:09 crc kubenswrapper[4923]: E0311 13:42:09.324645 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:42:10 crc kubenswrapper[4923]: I0311 13:42:10.364523 4923 scope.go:117] "RemoveContainer" containerID="8bf10a9ec7b294d63184d7c43bf1a0508fd8b54697b5139a5716aae8a4ba506c" Mar 11 13:42:10 crc kubenswrapper[4923]: I0311 13:42:10.364903 4923 scope.go:117] "RemoveContainer" containerID="4f323cc06c832423a935bd57211a21d27ae7dfdd630767d0c158db5de4fb1394" Mar 11 13:42:10 crc kubenswrapper[4923]: E0311 13:42:10.365213 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:42:11 crc kubenswrapper[4923]: I0311 13:42:11.444932 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:42:11 crc kubenswrapper[4923]: I0311 13:42:11.445910 4923 scope.go:117] "RemoveContainer" containerID="8bf10a9ec7b294d63184d7c43bf1a0508fd8b54697b5139a5716aae8a4ba506c" Mar 11 13:42:11 crc kubenswrapper[4923]: I0311 13:42:11.445942 4923 scope.go:117] "RemoveContainer" containerID="4f323cc06c832423a935bd57211a21d27ae7dfdd630767d0c158db5de4fb1394" Mar 11 13:42:11 crc kubenswrapper[4923]: E0311 13:42:11.446467 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:42:17 crc kubenswrapper[4923]: I0311 13:42:17.445428 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:42:17 crc kubenswrapper[4923]: I0311 13:42:17.446836 4923 scope.go:117] "RemoveContainer" containerID="8bf10a9ec7b294d63184d7c43bf1a0508fd8b54697b5139a5716aae8a4ba506c" Mar 11 13:42:17 crc kubenswrapper[4923]: I0311 13:42:17.446855 4923 scope.go:117] "RemoveContainer" containerID="4f323cc06c832423a935bd57211a21d27ae7dfdd630767d0c158db5de4fb1394" Mar 11 13:42:17 crc kubenswrapper[4923]: E0311 13:42:17.447220 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:42:30 crc kubenswrapper[4923]: I0311 13:42:30.022835 4923 scope.go:117] "RemoveContainer" containerID="8bf10a9ec7b294d63184d7c43bf1a0508fd8b54697b5139a5716aae8a4ba506c" Mar 11 13:42:30 crc kubenswrapper[4923]: I0311 13:42:30.023428 4923 scope.go:117] "RemoveContainer" containerID="4f323cc06c832423a935bd57211a21d27ae7dfdd630767d0c158db5de4fb1394" Mar 11 13:42:30 crc kubenswrapper[4923]: E0311 13:42:30.023798 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:42:33 crc kubenswrapper[4923]: I0311 13:42:33.995119 4923 scope.go:117] "RemoveContainer" containerID="9d946a60e806e217fa5f56b303307f58b580b8aff6f2cc08c8f6de3ffd1582a3" Mar 11 13:42:40 crc kubenswrapper[4923]: I0311 13:42:40.907326 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Mar 11 13:42:40 crc kubenswrapper[4923]: I0311 13:42:40.908026 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-0" podUID="fce6ff82-c436-41e1-bcef-f16413e39c71" containerName="cinder-scheduler" containerID="cri-o://196c02cd4bb53f4ecfc1a4200e37811d9001d1730c33f790c0922488994bf296" gracePeriod=30 Mar 11 13:42:40 crc kubenswrapper[4923]: I0311 13:42:40.908160 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-0" podUID="fce6ff82-c436-41e1-bcef-f16413e39c71" containerName="probe" containerID="cri-o://0457653217cc4be1bc77252e1a96fe9be0aa46fda288ff8562e108c88e27d142" gracePeriod=30 Mar 11 13:42:40 crc kubenswrapper[4923]: I0311 13:42:40.939586 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Mar 11 13:42:40 crc kubenswrapper[4923]: I0311 13:42:40.939962 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-0" podUID="d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" containerName="cinder-backup" containerID="cri-o://07d0c4907846fe6ec0f36a2b68ef7560e6274941e23d6b568ed6d0374c8ebad0" gracePeriod=30 Mar 11 13:42:40 crc kubenswrapper[4923]: I0311 13:42:40.940029 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-0" podUID="d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" containerName="probe" containerID="cri-o://882442eca39bb3e6fa387944706810da0348a93c5fd04094e8b1de3cecc8f208" gracePeriod=30 Mar 11 13:42:40 crc kubenswrapper[4923]: I0311 13:42:40.957742 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Mar 11 13:42:40 crc kubenswrapper[4923]: I0311 13:42:40.958070 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="811524db-4494-42cb-a404-25b40a5614f4" containerName="cinder-api-log" containerID="cri-o://a3bad500be86c9560cb179786f29d771d52a094d4367d0242c26c4822fdd0626" gracePeriod=30 Mar 11 13:42:40 crc kubenswrapper[4923]: I0311 13:42:40.958104 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="811524db-4494-42cb-a404-25b40a5614f4" containerName="cinder-api" containerID="cri-o://34ad1d334751d655aa2bd86e938ad9100f02f4969d25c60fd94512e6c86896ed" gracePeriod=30 Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.125697 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-gbxj9"] Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.135893 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-gbxj9"] Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.247516 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-bh8nh"] Mar 11 13:42:41 crc kubenswrapper[4923]: E0311 13:42:41.247813 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06883e63-9dba-48dd-a356-d990b50293ff" containerName="oc" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.247825 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="06883e63-9dba-48dd-a356-d990b50293ff" containerName="oc" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.247960 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="06883e63-9dba-48dd-a356-d990b50293ff" containerName="oc" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.248628 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.251405 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"combined-ca-bundle" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.257626 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-bh8nh"] Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.420762 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-config-data\") pod \"cinder-db-sync-bh8nh\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.420889 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/86738f89-b8e9-4cd5-806a-a54529ef1078-etc-machine-id\") pod \"cinder-db-sync-bh8nh\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.420932 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-combined-ca-bundle\") pod \"cinder-db-sync-bh8nh\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.420973 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-scripts\") pod \"cinder-db-sync-bh8nh\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.420994 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-db-sync-config-data\") pod \"cinder-db-sync-bh8nh\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.421185 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7v4l\" (UniqueName: \"kubernetes.io/projected/86738f89-b8e9-4cd5-806a-a54529ef1078-kube-api-access-p7v4l\") pod \"cinder-db-sync-bh8nh\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.522429 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/86738f89-b8e9-4cd5-806a-a54529ef1078-etc-machine-id\") pod \"cinder-db-sync-bh8nh\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.522519 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-combined-ca-bundle\") pod \"cinder-db-sync-bh8nh\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.522577 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-scripts\") pod \"cinder-db-sync-bh8nh\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.522583 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/86738f89-b8e9-4cd5-806a-a54529ef1078-etc-machine-id\") pod \"cinder-db-sync-bh8nh\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.522610 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-db-sync-config-data\") pod \"cinder-db-sync-bh8nh\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.522769 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7v4l\" (UniqueName: \"kubernetes.io/projected/86738f89-b8e9-4cd5-806a-a54529ef1078-kube-api-access-p7v4l\") pod \"cinder-db-sync-bh8nh\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.522848 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-config-data\") pod \"cinder-db-sync-bh8nh\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.530587 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-combined-ca-bundle\") pod \"cinder-db-sync-bh8nh\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.530675 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-db-sync-config-data\") pod \"cinder-db-sync-bh8nh\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.532657 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-scripts\") pod \"cinder-db-sync-bh8nh\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.539519 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-config-data\") pod \"cinder-db-sync-bh8nh\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.550006 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7v4l\" (UniqueName: \"kubernetes.io/projected/86738f89-b8e9-4cd5-806a-a54529ef1078-kube-api-access-p7v4l\") pod \"cinder-db-sync-bh8nh\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.565458 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.656054 4923 generic.go:334] "Generic (PLEG): container finished" podID="811524db-4494-42cb-a404-25b40a5614f4" containerID="a3bad500be86c9560cb179786f29d771d52a094d4367d0242c26c4822fdd0626" exitCode=143 Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.656135 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"811524db-4494-42cb-a404-25b40a5614f4","Type":"ContainerDied","Data":"a3bad500be86c9560cb179786f29d771d52a094d4367d0242c26c4822fdd0626"} Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.659278 4923 generic.go:334] "Generic (PLEG): container finished" podID="d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" containerID="882442eca39bb3e6fa387944706810da0348a93c5fd04094e8b1de3cecc8f208" exitCode=0 Mar 11 13:42:41 crc kubenswrapper[4923]: I0311 13:42:41.659323 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c","Type":"ContainerDied","Data":"882442eca39bb3e6fa387944706810da0348a93c5fd04094e8b1de3cecc8f208"} Mar 11 13:42:42 crc kubenswrapper[4923]: I0311 13:42:42.069575 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-bh8nh"] Mar 11 13:42:42 crc kubenswrapper[4923]: W0311 13:42:42.079402 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86738f89_b8e9_4cd5_806a_a54529ef1078.slice/crio-dc7f6756b9b9ad87004db39ba6aff08c789298ea6d0b8a709e8a6fe02b53e5fe WatchSource:0}: Error finding container dc7f6756b9b9ad87004db39ba6aff08c789298ea6d0b8a709e8a6fe02b53e5fe: Status 404 returned error can't find the container with id dc7f6756b9b9ad87004db39ba6aff08c789298ea6d0b8a709e8a6fe02b53e5fe Mar 11 13:42:42 crc kubenswrapper[4923]: I0311 13:42:42.673011 4923 generic.go:334] "Generic (PLEG): container finished" podID="fce6ff82-c436-41e1-bcef-f16413e39c71" containerID="0457653217cc4be1bc77252e1a96fe9be0aa46fda288ff8562e108c88e27d142" exitCode=0 Mar 11 13:42:42 crc kubenswrapper[4923]: I0311 13:42:42.673195 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"fce6ff82-c436-41e1-bcef-f16413e39c71","Type":"ContainerDied","Data":"0457653217cc4be1bc77252e1a96fe9be0aa46fda288ff8562e108c88e27d142"} Mar 11 13:42:42 crc kubenswrapper[4923]: I0311 13:42:42.675769 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" event={"ID":"86738f89-b8e9-4cd5-806a-a54529ef1078","Type":"ContainerStarted","Data":"1beb9859963659abd427eb2fe12c392da85f99762e7dcdccd8a85f9693702cd8"} Mar 11 13:42:42 crc kubenswrapper[4923]: I0311 13:42:42.675806 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" event={"ID":"86738f89-b8e9-4cd5-806a-a54529ef1078","Type":"ContainerStarted","Data":"dc7f6756b9b9ad87004db39ba6aff08c789298ea6d0b8a709e8a6fe02b53e5fe"} Mar 11 13:42:42 crc kubenswrapper[4923]: I0311 13:42:42.704701 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" podStartSLOduration=1.704682072 podStartE2EDuration="1.704682072s" podCreationTimestamp="2026-03-11 13:42:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:42:42.700277509 +0000 UTC m=+1210.382179263" watchObservedRunningTime="2026-03-11 13:42:42.704682072 +0000 UTC m=+1210.386583786" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.028035 4923 scope.go:117] "RemoveContainer" containerID="8bf10a9ec7b294d63184d7c43bf1a0508fd8b54697b5139a5716aae8a4ba506c" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.028065 4923 scope.go:117] "RemoveContainer" containerID="4f323cc06c832423a935bd57211a21d27ae7dfdd630767d0c158db5de4fb1394" Mar 11 13:42:43 crc kubenswrapper[4923]: E0311 13:42:43.028573 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.037462 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7361c58-e41e-42ab-9c24-d0e470ccea54" path="/var/lib/kubelet/pods/f7361c58-e41e-42ab-9c24-d0e470ccea54/volumes" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.168409 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.260217 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-scripts\") pod \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.262239 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-var-locks-cinder\") pod \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.262309 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" (UID: "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.262461 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-var-locks-brick\") pod \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.262509 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-etc-machine-id\") pod \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.262536 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-var-lib-cinder\") pod \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.262562 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-run\") pod \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.262613 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" (UID: "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.262665 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" (UID: "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.262721 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-config-data\") pod \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.262722 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" (UID: "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.262764 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-etc-nvme\") pod \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.262791 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-run" (OuterVolumeSpecName: "run") pod "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" (UID: "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.262802 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-lib-modules\") pod \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.262828 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" (UID: "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.262860 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" (UID: "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.262878 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-config-data-custom\") pod \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.262946 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5qc8\" (UniqueName: \"kubernetes.io/projected/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-kube-api-access-w5qc8\") pod \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.262991 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-etc-iscsi\") pod \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.263059 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-dev\") pod \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.263090 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-sys\") pod \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\" (UID: \"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c\") " Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.263259 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" (UID: "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.263292 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-dev" (OuterVolumeSpecName: "dev") pod "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" (UID: "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.263372 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-sys" (OuterVolumeSpecName: "sys") pod "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" (UID: "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.263843 4923 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-var-locks-brick\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.263875 4923 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.263895 4923 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.263963 4923 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-run\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.263985 4923 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-etc-nvme\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.264000 4923 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-lib-modules\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.264015 4923 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-etc-iscsi\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.264067 4923 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-dev\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.264087 4923 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-sys\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.264104 4923 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.266892 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-kube-api-access-w5qc8" (OuterVolumeSpecName: "kube-api-access-w5qc8") pod "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" (UID: "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c"). InnerVolumeSpecName "kube-api-access-w5qc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.266933 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" (UID: "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.284197 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-scripts" (OuterVolumeSpecName: "scripts") pod "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" (UID: "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.361046 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-config-data" (OuterVolumeSpecName: "config-data") pod "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" (UID: "d9b2b9ec-01a3-41a7-bf78-143b4be40a0c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.366163 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-config-data\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.366200 4923 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.366218 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5qc8\" (UniqueName: \"kubernetes.io/projected/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-kube-api-access-w5qc8\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.366233 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c-scripts\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.692391 4923 generic.go:334] "Generic (PLEG): container finished" podID="d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" containerID="07d0c4907846fe6ec0f36a2b68ef7560e6274941e23d6b568ed6d0374c8ebad0" exitCode=0 Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.692500 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.692482 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c","Type":"ContainerDied","Data":"07d0c4907846fe6ec0f36a2b68ef7560e6274941e23d6b568ed6d0374c8ebad0"} Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.692584 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"d9b2b9ec-01a3-41a7-bf78-143b4be40a0c","Type":"ContainerDied","Data":"30d12174232cc0bd0e2030e72644f8a2297fb7e9c2114fd6281b4caad66e07b9"} Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.692616 4923 scope.go:117] "RemoveContainer" containerID="882442eca39bb3e6fa387944706810da0348a93c5fd04094e8b1de3cecc8f208" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.734178 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.738997 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.739068 4923 scope.go:117] "RemoveContainer" containerID="07d0c4907846fe6ec0f36a2b68ef7560e6274941e23d6b568ed6d0374c8ebad0" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.773539 4923 scope.go:117] "RemoveContainer" containerID="882442eca39bb3e6fa387944706810da0348a93c5fd04094e8b1de3cecc8f208" Mar 11 13:42:43 crc kubenswrapper[4923]: E0311 13:42:43.774979 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"882442eca39bb3e6fa387944706810da0348a93c5fd04094e8b1de3cecc8f208\": container with ID starting with 882442eca39bb3e6fa387944706810da0348a93c5fd04094e8b1de3cecc8f208 not found: ID does not exist" containerID="882442eca39bb3e6fa387944706810da0348a93c5fd04094e8b1de3cecc8f208" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.775174 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"882442eca39bb3e6fa387944706810da0348a93c5fd04094e8b1de3cecc8f208"} err="failed to get container status \"882442eca39bb3e6fa387944706810da0348a93c5fd04094e8b1de3cecc8f208\": rpc error: code = NotFound desc = could not find container \"882442eca39bb3e6fa387944706810da0348a93c5fd04094e8b1de3cecc8f208\": container with ID starting with 882442eca39bb3e6fa387944706810da0348a93c5fd04094e8b1de3cecc8f208 not found: ID does not exist" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.775338 4923 scope.go:117] "RemoveContainer" containerID="07d0c4907846fe6ec0f36a2b68ef7560e6274941e23d6b568ed6d0374c8ebad0" Mar 11 13:42:43 crc kubenswrapper[4923]: E0311 13:42:43.775948 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07d0c4907846fe6ec0f36a2b68ef7560e6274941e23d6b568ed6d0374c8ebad0\": container with ID starting with 07d0c4907846fe6ec0f36a2b68ef7560e6274941e23d6b568ed6d0374c8ebad0 not found: ID does not exist" containerID="07d0c4907846fe6ec0f36a2b68ef7560e6274941e23d6b568ed6d0374c8ebad0" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.776141 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07d0c4907846fe6ec0f36a2b68ef7560e6274941e23d6b568ed6d0374c8ebad0"} err="failed to get container status \"07d0c4907846fe6ec0f36a2b68ef7560e6274941e23d6b568ed6d0374c8ebad0\": rpc error: code = NotFound desc = could not find container \"07d0c4907846fe6ec0f36a2b68ef7560e6274941e23d6b568ed6d0374c8ebad0\": container with ID starting with 07d0c4907846fe6ec0f36a2b68ef7560e6274941e23d6b568ed6d0374c8ebad0 not found: ID does not exist" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.788655 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Mar 11 13:42:43 crc kubenswrapper[4923]: E0311 13:42:43.788918 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" containerName="cinder-backup" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.788934 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" containerName="cinder-backup" Mar 11 13:42:43 crc kubenswrapper[4923]: E0311 13:42:43.788948 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" containerName="probe" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.788954 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" containerName="probe" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.789088 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" containerName="probe" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.789099 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" containerName="cinder-backup" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.789747 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.795557 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-backup-config-data" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.860185 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.978816 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-run\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.978863 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-lib-modules\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.978894 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f87edc5e-c658-4ede-a5a2-157f3b97ee80-scripts\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.978912 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f87edc5e-c658-4ede-a5a2-157f3b97ee80-config-data-custom\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.978930 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-sys\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.978946 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kz2d\" (UniqueName: \"kubernetes.io/projected/f87edc5e-c658-4ede-a5a2-157f3b97ee80-kube-api-access-4kz2d\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.978967 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-etc-nvme\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.978989 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.979022 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-dev\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.979040 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.979060 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.979079 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.979103 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f87edc5e-c658-4ede-a5a2-157f3b97ee80-config-data\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:43 crc kubenswrapper[4923]: I0311 13:42:43.979116 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.081059 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.081171 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-dev\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.081216 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.081258 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.081269 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.081296 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.081390 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.081416 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f87edc5e-c658-4ede-a5a2-157f3b97ee80-config-data\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.081445 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.081509 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.081527 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-run\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.081554 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.081561 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-lib-modules\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.081590 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-lib-modules\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.081637 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f87edc5e-c658-4ede-a5a2-157f3b97ee80-scripts\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.081671 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f87edc5e-c658-4ede-a5a2-157f3b97ee80-config-data-custom\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.081706 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-sys\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.081739 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kz2d\" (UniqueName: \"kubernetes.io/projected/f87edc5e-c658-4ede-a5a2-157f3b97ee80-kube-api-access-4kz2d\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.081778 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-etc-nvme\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.081918 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-etc-nvme\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.082129 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-sys\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.081447 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-dev\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.083660 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-run\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.083854 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.087150 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f87edc5e-c658-4ede-a5a2-157f3b97ee80-config-data\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.087475 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f87edc5e-c658-4ede-a5a2-157f3b97ee80-config-data-custom\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.087748 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f87edc5e-c658-4ede-a5a2-157f3b97ee80-scripts\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.117022 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kz2d\" (UniqueName: \"kubernetes.io/projected/f87edc5e-c658-4ede-a5a2-157f3b97ee80-kube-api-access-4kz2d\") pod \"cinder-backup-0\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.127683 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.476716 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.569459 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Mar 11 13:42:44 crc kubenswrapper[4923]: W0311 13:42:44.575977 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf87edc5e_c658_4ede_a5a2_157f3b97ee80.slice/crio-73e2bba53e451c81665f0b82c801fe1d9becc3d3bba38f4971178daa37bb2706 WatchSource:0}: Error finding container 73e2bba53e451c81665f0b82c801fe1d9becc3d3bba38f4971178daa37bb2706: Status 404 returned error can't find the container with id 73e2bba53e451c81665f0b82c801fe1d9becc3d3bba38f4971178daa37bb2706 Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.590361 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/811524db-4494-42cb-a404-25b40a5614f4-scripts\") pod \"811524db-4494-42cb-a404-25b40a5614f4\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.590463 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/811524db-4494-42cb-a404-25b40a5614f4-config-data\") pod \"811524db-4494-42cb-a404-25b40a5614f4\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.590496 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrn27\" (UniqueName: \"kubernetes.io/projected/811524db-4494-42cb-a404-25b40a5614f4-kube-api-access-mrn27\") pod \"811524db-4494-42cb-a404-25b40a5614f4\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.590550 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/811524db-4494-42cb-a404-25b40a5614f4-etc-machine-id\") pod \"811524db-4494-42cb-a404-25b40a5614f4\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.590617 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/811524db-4494-42cb-a404-25b40a5614f4-logs\") pod \"811524db-4494-42cb-a404-25b40a5614f4\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.590643 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/811524db-4494-42cb-a404-25b40a5614f4-config-data-custom\") pod \"811524db-4494-42cb-a404-25b40a5614f4\" (UID: \"811524db-4494-42cb-a404-25b40a5614f4\") " Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.590990 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/811524db-4494-42cb-a404-25b40a5614f4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "811524db-4494-42cb-a404-25b40a5614f4" (UID: "811524db-4494-42cb-a404-25b40a5614f4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.591508 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/811524db-4494-42cb-a404-25b40a5614f4-logs" (OuterVolumeSpecName: "logs") pod "811524db-4494-42cb-a404-25b40a5614f4" (UID: "811524db-4494-42cb-a404-25b40a5614f4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.593678 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/811524db-4494-42cb-a404-25b40a5614f4-kube-api-access-mrn27" (OuterVolumeSpecName: "kube-api-access-mrn27") pod "811524db-4494-42cb-a404-25b40a5614f4" (UID: "811524db-4494-42cb-a404-25b40a5614f4"). InnerVolumeSpecName "kube-api-access-mrn27". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.594228 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/811524db-4494-42cb-a404-25b40a5614f4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "811524db-4494-42cb-a404-25b40a5614f4" (UID: "811524db-4494-42cb-a404-25b40a5614f4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.594934 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/811524db-4494-42cb-a404-25b40a5614f4-scripts" (OuterVolumeSpecName: "scripts") pod "811524db-4494-42cb-a404-25b40a5614f4" (UID: "811524db-4494-42cb-a404-25b40a5614f4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.636844 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/811524db-4494-42cb-a404-25b40a5614f4-config-data" (OuterVolumeSpecName: "config-data") pod "811524db-4494-42cb-a404-25b40a5614f4" (UID: "811524db-4494-42cb-a404-25b40a5614f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.692273 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrn27\" (UniqueName: \"kubernetes.io/projected/811524db-4494-42cb-a404-25b40a5614f4-kube-api-access-mrn27\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.692304 4923 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/811524db-4494-42cb-a404-25b40a5614f4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.692313 4923 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/811524db-4494-42cb-a404-25b40a5614f4-logs\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.692322 4923 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/811524db-4494-42cb-a404-25b40a5614f4-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.692329 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/811524db-4494-42cb-a404-25b40a5614f4-scripts\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.692344 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/811524db-4494-42cb-a404-25b40a5614f4-config-data\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.699325 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"f87edc5e-c658-4ede-a5a2-157f3b97ee80","Type":"ContainerStarted","Data":"73e2bba53e451c81665f0b82c801fe1d9becc3d3bba38f4971178daa37bb2706"} Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.700747 4923 generic.go:334] "Generic (PLEG): container finished" podID="811524db-4494-42cb-a404-25b40a5614f4" containerID="34ad1d334751d655aa2bd86e938ad9100f02f4969d25c60fd94512e6c86896ed" exitCode=0 Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.700797 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.700824 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"811524db-4494-42cb-a404-25b40a5614f4","Type":"ContainerDied","Data":"34ad1d334751d655aa2bd86e938ad9100f02f4969d25c60fd94512e6c86896ed"} Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.700910 4923 scope.go:117] "RemoveContainer" containerID="34ad1d334751d655aa2bd86e938ad9100f02f4969d25c60fd94512e6c86896ed" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.700908 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"811524db-4494-42cb-a404-25b40a5614f4","Type":"ContainerDied","Data":"e56961fcfe8fc2210d49892cf2e92d85fbabc1b4ca06b36d9d1058b305dcda40"} Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.703474 4923 generic.go:334] "Generic (PLEG): container finished" podID="86738f89-b8e9-4cd5-806a-a54529ef1078" containerID="1beb9859963659abd427eb2fe12c392da85f99762e7dcdccd8a85f9693702cd8" exitCode=0 Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.703500 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" event={"ID":"86738f89-b8e9-4cd5-806a-a54529ef1078","Type":"ContainerDied","Data":"1beb9859963659abd427eb2fe12c392da85f99762e7dcdccd8a85f9693702cd8"} Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.755665 4923 scope.go:117] "RemoveContainer" containerID="a3bad500be86c9560cb179786f29d771d52a094d4367d0242c26c4822fdd0626" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.764861 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.773951 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.798926 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Mar 11 13:42:44 crc kubenswrapper[4923]: E0311 13:42:44.799214 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="811524db-4494-42cb-a404-25b40a5614f4" containerName="cinder-api" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.799232 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="811524db-4494-42cb-a404-25b40a5614f4" containerName="cinder-api" Mar 11 13:42:44 crc kubenswrapper[4923]: E0311 13:42:44.799271 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="811524db-4494-42cb-a404-25b40a5614f4" containerName="cinder-api-log" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.799283 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="811524db-4494-42cb-a404-25b40a5614f4" containerName="cinder-api-log" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.799432 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="811524db-4494-42cb-a404-25b40a5614f4" containerName="cinder-api" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.799468 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="811524db-4494-42cb-a404-25b40a5614f4" containerName="cinder-api-log" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.805766 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.805897 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.808291 4923 scope.go:117] "RemoveContainer" containerID="34ad1d334751d655aa2bd86e938ad9100f02f4969d25c60fd94512e6c86896ed" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.808537 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-api-config-data" Mar 11 13:42:44 crc kubenswrapper[4923]: E0311 13:42:44.808657 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34ad1d334751d655aa2bd86e938ad9100f02f4969d25c60fd94512e6c86896ed\": container with ID starting with 34ad1d334751d655aa2bd86e938ad9100f02f4969d25c60fd94512e6c86896ed not found: ID does not exist" containerID="34ad1d334751d655aa2bd86e938ad9100f02f4969d25c60fd94512e6c86896ed" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.808685 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34ad1d334751d655aa2bd86e938ad9100f02f4969d25c60fd94512e6c86896ed"} err="failed to get container status \"34ad1d334751d655aa2bd86e938ad9100f02f4969d25c60fd94512e6c86896ed\": rpc error: code = NotFound desc = could not find container \"34ad1d334751d655aa2bd86e938ad9100f02f4969d25c60fd94512e6c86896ed\": container with ID starting with 34ad1d334751d655aa2bd86e938ad9100f02f4969d25c60fd94512e6c86896ed not found: ID does not exist" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.808710 4923 scope.go:117] "RemoveContainer" containerID="a3bad500be86c9560cb179786f29d771d52a094d4367d0242c26c4822fdd0626" Mar 11 13:42:44 crc kubenswrapper[4923]: E0311 13:42:44.809011 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3bad500be86c9560cb179786f29d771d52a094d4367d0242c26c4822fdd0626\": container with ID starting with a3bad500be86c9560cb179786f29d771d52a094d4367d0242c26c4822fdd0626 not found: ID does not exist" containerID="a3bad500be86c9560cb179786f29d771d52a094d4367d0242c26c4822fdd0626" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.809054 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3bad500be86c9560cb179786f29d771d52a094d4367d0242c26c4822fdd0626"} err="failed to get container status \"a3bad500be86c9560cb179786f29d771d52a094d4367d0242c26c4822fdd0626\": rpc error: code = NotFound desc = could not find container \"a3bad500be86c9560cb179786f29d771d52a094d4367d0242c26c4822fdd0626\": container with ID starting with a3bad500be86c9560cb179786f29d771d52a094d4367d0242c26c4822fdd0626 not found: ID does not exist" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.906322 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b138d16e-9b69-4a49-8182-53bcb9346977-config-data\") pod \"cinder-api-0\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.906668 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b138d16e-9b69-4a49-8182-53bcb9346977-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.906690 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b138d16e-9b69-4a49-8182-53bcb9346977-scripts\") pod \"cinder-api-0\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.906908 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b138d16e-9b69-4a49-8182-53bcb9346977-config-data-custom\") pod \"cinder-api-0\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.906976 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b138d16e-9b69-4a49-8182-53bcb9346977-logs\") pod \"cinder-api-0\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:44 crc kubenswrapper[4923]: I0311 13:42:44.907036 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7sgb\" (UniqueName: \"kubernetes.io/projected/b138d16e-9b69-4a49-8182-53bcb9346977-kube-api-access-c7sgb\") pod \"cinder-api-0\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:45 crc kubenswrapper[4923]: I0311 13:42:45.008705 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b138d16e-9b69-4a49-8182-53bcb9346977-config-data-custom\") pod \"cinder-api-0\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:45 crc kubenswrapper[4923]: I0311 13:42:45.008823 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b138d16e-9b69-4a49-8182-53bcb9346977-logs\") pod \"cinder-api-0\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:45 crc kubenswrapper[4923]: I0311 13:42:45.008898 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7sgb\" (UniqueName: \"kubernetes.io/projected/b138d16e-9b69-4a49-8182-53bcb9346977-kube-api-access-c7sgb\") pod \"cinder-api-0\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:45 crc kubenswrapper[4923]: I0311 13:42:45.008986 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b138d16e-9b69-4a49-8182-53bcb9346977-config-data\") pod \"cinder-api-0\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:45 crc kubenswrapper[4923]: I0311 13:42:45.009182 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b138d16e-9b69-4a49-8182-53bcb9346977-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:45 crc kubenswrapper[4923]: I0311 13:42:45.009230 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b138d16e-9b69-4a49-8182-53bcb9346977-scripts\") pod \"cinder-api-0\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:45 crc kubenswrapper[4923]: I0311 13:42:45.009806 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b138d16e-9b69-4a49-8182-53bcb9346977-logs\") pod \"cinder-api-0\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:45 crc kubenswrapper[4923]: I0311 13:42:45.009885 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b138d16e-9b69-4a49-8182-53bcb9346977-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:45 crc kubenswrapper[4923]: I0311 13:42:45.015807 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b138d16e-9b69-4a49-8182-53bcb9346977-config-data-custom\") pod \"cinder-api-0\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:45 crc kubenswrapper[4923]: I0311 13:42:45.016166 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b138d16e-9b69-4a49-8182-53bcb9346977-scripts\") pod \"cinder-api-0\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:45 crc kubenswrapper[4923]: I0311 13:42:45.017318 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b138d16e-9b69-4a49-8182-53bcb9346977-config-data\") pod \"cinder-api-0\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:45 crc kubenswrapper[4923]: I0311 13:42:45.030018 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7sgb\" (UniqueName: \"kubernetes.io/projected/b138d16e-9b69-4a49-8182-53bcb9346977-kube-api-access-c7sgb\") pod \"cinder-api-0\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:45 crc kubenswrapper[4923]: I0311 13:42:45.039607 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="811524db-4494-42cb-a404-25b40a5614f4" path="/var/lib/kubelet/pods/811524db-4494-42cb-a404-25b40a5614f4/volumes" Mar 11 13:42:45 crc kubenswrapper[4923]: I0311 13:42:45.041802 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9b2b9ec-01a3-41a7-bf78-143b4be40a0c" path="/var/lib/kubelet/pods/d9b2b9ec-01a3-41a7-bf78-143b4be40a0c/volumes" Mar 11 13:42:45 crc kubenswrapper[4923]: I0311 13:42:45.138665 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:45 crc kubenswrapper[4923]: I0311 13:42:45.459903 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Mar 11 13:42:45 crc kubenswrapper[4923]: I0311 13:42:45.731307 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"f87edc5e-c658-4ede-a5a2-157f3b97ee80","Type":"ContainerStarted","Data":"af06caaf9c4e3c325cb65dbb5745eb7e0c7abc06255512446c9595be987ac3c3"} Mar 11 13:42:45 crc kubenswrapper[4923]: I0311 13:42:45.731415 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"f87edc5e-c658-4ede-a5a2-157f3b97ee80","Type":"ContainerStarted","Data":"898fd024cb5a8479908125b72c3ba3ad34db8ecf31d0d64ed129377988f9982b"} Mar 11 13:42:45 crc kubenswrapper[4923]: I0311 13:42:45.737456 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"b138d16e-9b69-4a49-8182-53bcb9346977","Type":"ContainerStarted","Data":"0eca812f9b38993489557623fe0e79a970b63d75630476385053c4bd6a3843e0"} Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.001146 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.019323 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-backup-0" podStartSLOduration=3.019301636 podStartE2EDuration="3.019301636s" podCreationTimestamp="2026-03-11 13:42:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:42:45.763839716 +0000 UTC m=+1213.445741470" watchObservedRunningTime="2026-03-11 13:42:46.019301636 +0000 UTC m=+1213.701203350" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.130393 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-config-data\") pod \"86738f89-b8e9-4cd5-806a-a54529ef1078\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.130504 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-combined-ca-bundle\") pod \"86738f89-b8e9-4cd5-806a-a54529ef1078\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.130539 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-scripts\") pod \"86738f89-b8e9-4cd5-806a-a54529ef1078\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.130570 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/86738f89-b8e9-4cd5-806a-a54529ef1078-etc-machine-id\") pod \"86738f89-b8e9-4cd5-806a-a54529ef1078\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.130633 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7v4l\" (UniqueName: \"kubernetes.io/projected/86738f89-b8e9-4cd5-806a-a54529ef1078-kube-api-access-p7v4l\") pod \"86738f89-b8e9-4cd5-806a-a54529ef1078\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.130675 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-db-sync-config-data\") pod \"86738f89-b8e9-4cd5-806a-a54529ef1078\" (UID: \"86738f89-b8e9-4cd5-806a-a54529ef1078\") " Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.131724 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86738f89-b8e9-4cd5-806a-a54529ef1078-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "86738f89-b8e9-4cd5-806a-a54529ef1078" (UID: "86738f89-b8e9-4cd5-806a-a54529ef1078"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.134590 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86738f89-b8e9-4cd5-806a-a54529ef1078-kube-api-access-p7v4l" (OuterVolumeSpecName: "kube-api-access-p7v4l") pod "86738f89-b8e9-4cd5-806a-a54529ef1078" (UID: "86738f89-b8e9-4cd5-806a-a54529ef1078"). InnerVolumeSpecName "kube-api-access-p7v4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.136406 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-scripts" (OuterVolumeSpecName: "scripts") pod "86738f89-b8e9-4cd5-806a-a54529ef1078" (UID: "86738f89-b8e9-4cd5-806a-a54529ef1078"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.137300 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "86738f89-b8e9-4cd5-806a-a54529ef1078" (UID: "86738f89-b8e9-4cd5-806a-a54529ef1078"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.153406 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "86738f89-b8e9-4cd5-806a-a54529ef1078" (UID: "86738f89-b8e9-4cd5-806a-a54529ef1078"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.178248 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-config-data" (OuterVolumeSpecName: "config-data") pod "86738f89-b8e9-4cd5-806a-a54529ef1078" (UID: "86738f89-b8e9-4cd5-806a-a54529ef1078"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.233132 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7v4l\" (UniqueName: \"kubernetes.io/projected/86738f89-b8e9-4cd5-806a-a54529ef1078-kube-api-access-p7v4l\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.233174 4923 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.233188 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-config-data\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.233198 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.233209 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/86738f89-b8e9-4cd5-806a-a54529ef1078-scripts\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.233219 4923 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/86738f89-b8e9-4cd5-806a-a54529ef1078-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.391343 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.536299 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fce6ff82-c436-41e1-bcef-f16413e39c71-etc-machine-id\") pod \"fce6ff82-c436-41e1-bcef-f16413e39c71\" (UID: \"fce6ff82-c436-41e1-bcef-f16413e39c71\") " Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.536384 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgp9m\" (UniqueName: \"kubernetes.io/projected/fce6ff82-c436-41e1-bcef-f16413e39c71-kube-api-access-wgp9m\") pod \"fce6ff82-c436-41e1-bcef-f16413e39c71\" (UID: \"fce6ff82-c436-41e1-bcef-f16413e39c71\") " Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.536457 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fce6ff82-c436-41e1-bcef-f16413e39c71-scripts\") pod \"fce6ff82-c436-41e1-bcef-f16413e39c71\" (UID: \"fce6ff82-c436-41e1-bcef-f16413e39c71\") " Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.536498 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fce6ff82-c436-41e1-bcef-f16413e39c71-config-data\") pod \"fce6ff82-c436-41e1-bcef-f16413e39c71\" (UID: \"fce6ff82-c436-41e1-bcef-f16413e39c71\") " Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.536499 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fce6ff82-c436-41e1-bcef-f16413e39c71-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "fce6ff82-c436-41e1-bcef-f16413e39c71" (UID: "fce6ff82-c436-41e1-bcef-f16413e39c71"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.536525 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fce6ff82-c436-41e1-bcef-f16413e39c71-config-data-custom\") pod \"fce6ff82-c436-41e1-bcef-f16413e39c71\" (UID: \"fce6ff82-c436-41e1-bcef-f16413e39c71\") " Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.537821 4923 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fce6ff82-c436-41e1-bcef-f16413e39c71-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.540635 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fce6ff82-c436-41e1-bcef-f16413e39c71-scripts" (OuterVolumeSpecName: "scripts") pod "fce6ff82-c436-41e1-bcef-f16413e39c71" (UID: "fce6ff82-c436-41e1-bcef-f16413e39c71"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.540867 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fce6ff82-c436-41e1-bcef-f16413e39c71-kube-api-access-wgp9m" (OuterVolumeSpecName: "kube-api-access-wgp9m") pod "fce6ff82-c436-41e1-bcef-f16413e39c71" (UID: "fce6ff82-c436-41e1-bcef-f16413e39c71"). InnerVolumeSpecName "kube-api-access-wgp9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.543204 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fce6ff82-c436-41e1-bcef-f16413e39c71-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fce6ff82-c436-41e1-bcef-f16413e39c71" (UID: "fce6ff82-c436-41e1-bcef-f16413e39c71"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.597329 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fce6ff82-c436-41e1-bcef-f16413e39c71-config-data" (OuterVolumeSpecName: "config-data") pod "fce6ff82-c436-41e1-bcef-f16413e39c71" (UID: "fce6ff82-c436-41e1-bcef-f16413e39c71"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.639187 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fce6ff82-c436-41e1-bcef-f16413e39c71-scripts\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.639433 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fce6ff82-c436-41e1-bcef-f16413e39c71-config-data\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.639570 4923 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fce6ff82-c436-41e1-bcef-f16413e39c71-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.639701 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgp9m\" (UniqueName: \"kubernetes.io/projected/fce6ff82-c436-41e1-bcef-f16413e39c71-kube-api-access-wgp9m\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.762868 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" event={"ID":"86738f89-b8e9-4cd5-806a-a54529ef1078","Type":"ContainerDied","Data":"dc7f6756b9b9ad87004db39ba6aff08c789298ea6d0b8a709e8a6fe02b53e5fe"} Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.762921 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc7f6756b9b9ad87004db39ba6aff08c789298ea6d0b8a709e8a6fe02b53e5fe" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.762997 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-db-sync-bh8nh" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.773661 4923 generic.go:334] "Generic (PLEG): container finished" podID="fce6ff82-c436-41e1-bcef-f16413e39c71" containerID="196c02cd4bb53f4ecfc1a4200e37811d9001d1730c33f790c0922488994bf296" exitCode=0 Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.774106 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"fce6ff82-c436-41e1-bcef-f16413e39c71","Type":"ContainerDied","Data":"196c02cd4bb53f4ecfc1a4200e37811d9001d1730c33f790c0922488994bf296"} Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.774270 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"fce6ff82-c436-41e1-bcef-f16413e39c71","Type":"ContainerDied","Data":"dd5b8d7960504e138033dd1e12938f82ee1bb160331c04dacf527003aaf53cb9"} Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.774429 4923 scope.go:117] "RemoveContainer" containerID="0457653217cc4be1bc77252e1a96fe9be0aa46fda288ff8562e108c88e27d142" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.774807 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.788203 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"b138d16e-9b69-4a49-8182-53bcb9346977","Type":"ContainerStarted","Data":"c31b29598b06dcee0d8310749c3cf3bdb877be2865fba41550baadc936f9b3f8"} Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.789836 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.789853 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"b138d16e-9b69-4a49-8182-53bcb9346977","Type":"ContainerStarted","Data":"79905bbb8aebd892eadaf02ecc3305cf7393b8e1caf4211e0dbf6f70950da97e"} Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.807306 4923 scope.go:117] "RemoveContainer" containerID="196c02cd4bb53f4ecfc1a4200e37811d9001d1730c33f790c0922488994bf296" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.829381 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-0" podStartSLOduration=2.829363866 podStartE2EDuration="2.829363866s" podCreationTimestamp="2026-03-11 13:42:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:42:46.817715495 +0000 UTC m=+1214.499617209" watchObservedRunningTime="2026-03-11 13:42:46.829363866 +0000 UTC m=+1214.511265580" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.847280 4923 scope.go:117] "RemoveContainer" containerID="0457653217cc4be1bc77252e1a96fe9be0aa46fda288ff8562e108c88e27d142" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.852300 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Mar 11 13:42:46 crc kubenswrapper[4923]: E0311 13:42:46.854817 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0457653217cc4be1bc77252e1a96fe9be0aa46fda288ff8562e108c88e27d142\": container with ID starting with 0457653217cc4be1bc77252e1a96fe9be0aa46fda288ff8562e108c88e27d142 not found: ID does not exist" containerID="0457653217cc4be1bc77252e1a96fe9be0aa46fda288ff8562e108c88e27d142" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.854869 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0457653217cc4be1bc77252e1a96fe9be0aa46fda288ff8562e108c88e27d142"} err="failed to get container status \"0457653217cc4be1bc77252e1a96fe9be0aa46fda288ff8562e108c88e27d142\": rpc error: code = NotFound desc = could not find container \"0457653217cc4be1bc77252e1a96fe9be0aa46fda288ff8562e108c88e27d142\": container with ID starting with 0457653217cc4be1bc77252e1a96fe9be0aa46fda288ff8562e108c88e27d142 not found: ID does not exist" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.854907 4923 scope.go:117] "RemoveContainer" containerID="196c02cd4bb53f4ecfc1a4200e37811d9001d1730c33f790c0922488994bf296" Mar 11 13:42:46 crc kubenswrapper[4923]: E0311 13:42:46.858728 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"196c02cd4bb53f4ecfc1a4200e37811d9001d1730c33f790c0922488994bf296\": container with ID starting with 196c02cd4bb53f4ecfc1a4200e37811d9001d1730c33f790c0922488994bf296 not found: ID does not exist" containerID="196c02cd4bb53f4ecfc1a4200e37811d9001d1730c33f790c0922488994bf296" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.858757 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"196c02cd4bb53f4ecfc1a4200e37811d9001d1730c33f790c0922488994bf296"} err="failed to get container status \"196c02cd4bb53f4ecfc1a4200e37811d9001d1730c33f790c0922488994bf296\": rpc error: code = NotFound desc = could not find container \"196c02cd4bb53f4ecfc1a4200e37811d9001d1730c33f790c0922488994bf296\": container with ID starting with 196c02cd4bb53f4ecfc1a4200e37811d9001d1730c33f790c0922488994bf296 not found: ID does not exist" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.877485 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.886311 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Mar 11 13:42:46 crc kubenswrapper[4923]: E0311 13:42:46.886694 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fce6ff82-c436-41e1-bcef-f16413e39c71" containerName="cinder-scheduler" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.886716 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="fce6ff82-c436-41e1-bcef-f16413e39c71" containerName="cinder-scheduler" Mar 11 13:42:46 crc kubenswrapper[4923]: E0311 13:42:46.886729 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fce6ff82-c436-41e1-bcef-f16413e39c71" containerName="probe" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.886737 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="fce6ff82-c436-41e1-bcef-f16413e39c71" containerName="probe" Mar 11 13:42:46 crc kubenswrapper[4923]: E0311 13:42:46.886745 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86738f89-b8e9-4cd5-806a-a54529ef1078" containerName="cinder-db-sync" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.886758 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="86738f89-b8e9-4cd5-806a-a54529ef1078" containerName="cinder-db-sync" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.886866 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="fce6ff82-c436-41e1-bcef-f16413e39c71" containerName="cinder-scheduler" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.886883 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="86738f89-b8e9-4cd5-806a-a54529ef1078" containerName="cinder-db-sync" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.886900 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="fce6ff82-c436-41e1-bcef-f16413e39c71" containerName="probe" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.887591 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.892659 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scheduler-config-data" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.894210 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.942439 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h64tg\" (UniqueName: \"kubernetes.io/projected/97b85261-6521-4ca2-b7e9-b6fc59e63971-kube-api-access-h64tg\") pod \"cinder-scheduler-0\" (UID: \"97b85261-6521-4ca2-b7e9-b6fc59e63971\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.942813 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/97b85261-6521-4ca2-b7e9-b6fc59e63971-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"97b85261-6521-4ca2-b7e9-b6fc59e63971\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.942841 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97b85261-6521-4ca2-b7e9-b6fc59e63971-scripts\") pod \"cinder-scheduler-0\" (UID: \"97b85261-6521-4ca2-b7e9-b6fc59e63971\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.942860 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97b85261-6521-4ca2-b7e9-b6fc59e63971-config-data\") pod \"cinder-scheduler-0\" (UID: \"97b85261-6521-4ca2-b7e9-b6fc59e63971\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:42:46 crc kubenswrapper[4923]: I0311 13:42:46.942937 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/97b85261-6521-4ca2-b7e9-b6fc59e63971-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"97b85261-6521-4ca2-b7e9-b6fc59e63971\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:42:47 crc kubenswrapper[4923]: I0311 13:42:47.011485 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Mar 11 13:42:47 crc kubenswrapper[4923]: I0311 13:42:47.032746 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fce6ff82-c436-41e1-bcef-f16413e39c71" path="/var/lib/kubelet/pods/fce6ff82-c436-41e1-bcef-f16413e39c71/volumes" Mar 11 13:42:47 crc kubenswrapper[4923]: I0311 13:42:47.044184 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/97b85261-6521-4ca2-b7e9-b6fc59e63971-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"97b85261-6521-4ca2-b7e9-b6fc59e63971\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:42:47 crc kubenswrapper[4923]: I0311 13:42:47.044437 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h64tg\" (UniqueName: \"kubernetes.io/projected/97b85261-6521-4ca2-b7e9-b6fc59e63971-kube-api-access-h64tg\") pod \"cinder-scheduler-0\" (UID: \"97b85261-6521-4ca2-b7e9-b6fc59e63971\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:42:47 crc kubenswrapper[4923]: I0311 13:42:47.044537 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/97b85261-6521-4ca2-b7e9-b6fc59e63971-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"97b85261-6521-4ca2-b7e9-b6fc59e63971\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:42:47 crc kubenswrapper[4923]: I0311 13:42:47.044616 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97b85261-6521-4ca2-b7e9-b6fc59e63971-scripts\") pod \"cinder-scheduler-0\" (UID: \"97b85261-6521-4ca2-b7e9-b6fc59e63971\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:42:47 crc kubenswrapper[4923]: I0311 13:42:47.044699 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97b85261-6521-4ca2-b7e9-b6fc59e63971-config-data\") pod \"cinder-scheduler-0\" (UID: \"97b85261-6521-4ca2-b7e9-b6fc59e63971\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:42:47 crc kubenswrapper[4923]: I0311 13:42:47.046612 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/97b85261-6521-4ca2-b7e9-b6fc59e63971-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"97b85261-6521-4ca2-b7e9-b6fc59e63971\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:42:47 crc kubenswrapper[4923]: I0311 13:42:47.049526 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97b85261-6521-4ca2-b7e9-b6fc59e63971-scripts\") pod \"cinder-scheduler-0\" (UID: \"97b85261-6521-4ca2-b7e9-b6fc59e63971\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:42:47 crc kubenswrapper[4923]: I0311 13:42:47.049922 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97b85261-6521-4ca2-b7e9-b6fc59e63971-config-data\") pod \"cinder-scheduler-0\" (UID: \"97b85261-6521-4ca2-b7e9-b6fc59e63971\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:42:47 crc kubenswrapper[4923]: I0311 13:42:47.050870 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/97b85261-6521-4ca2-b7e9-b6fc59e63971-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"97b85261-6521-4ca2-b7e9-b6fc59e63971\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:42:47 crc kubenswrapper[4923]: I0311 13:42:47.076179 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h64tg\" (UniqueName: \"kubernetes.io/projected/97b85261-6521-4ca2-b7e9-b6fc59e63971-kube-api-access-h64tg\") pod \"cinder-scheduler-0\" (UID: \"97b85261-6521-4ca2-b7e9-b6fc59e63971\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:42:47 crc kubenswrapper[4923]: I0311 13:42:47.224680 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:42:47 crc kubenswrapper[4923]: I0311 13:42:47.412370 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Mar 11 13:42:47 crc kubenswrapper[4923]: W0311 13:42:47.430695 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97b85261_6521_4ca2_b7e9_b6fc59e63971.slice/crio-3b88a725c03ea4013d509461c020d6da59240b457be505dd9003478a180f27b8 WatchSource:0}: Error finding container 3b88a725c03ea4013d509461c020d6da59240b457be505dd9003478a180f27b8: Status 404 returned error can't find the container with id 3b88a725c03ea4013d509461c020d6da59240b457be505dd9003478a180f27b8 Mar 11 13:42:47 crc kubenswrapper[4923]: I0311 13:42:47.801532 4923 generic.go:334] "Generic (PLEG): container finished" podID="b138d16e-9b69-4a49-8182-53bcb9346977" containerID="c31b29598b06dcee0d8310749c3cf3bdb877be2865fba41550baadc936f9b3f8" exitCode=1 Mar 11 13:42:47 crc kubenswrapper[4923]: I0311 13:42:47.801606 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"b138d16e-9b69-4a49-8182-53bcb9346977","Type":"ContainerDied","Data":"c31b29598b06dcee0d8310749c3cf3bdb877be2865fba41550baadc936f9b3f8"} Mar 11 13:42:47 crc kubenswrapper[4923]: I0311 13:42:47.802065 4923 scope.go:117] "RemoveContainer" containerID="c31b29598b06dcee0d8310749c3cf3bdb877be2865fba41550baadc936f9b3f8" Mar 11 13:42:47 crc kubenswrapper[4923]: I0311 13:42:47.802829 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"97b85261-6521-4ca2-b7e9-b6fc59e63971","Type":"ContainerStarted","Data":"3b88a725c03ea4013d509461c020d6da59240b457be505dd9003478a180f27b8"} Mar 11 13:42:48 crc kubenswrapper[4923]: I0311 13:42:48.817416 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"97b85261-6521-4ca2-b7e9-b6fc59e63971","Type":"ContainerStarted","Data":"4824ba162fa478a80e5086b293c197316a3e376385333aa14602521ad8e30f25"} Mar 11 13:42:48 crc kubenswrapper[4923]: I0311 13:42:48.818533 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"97b85261-6521-4ca2-b7e9-b6fc59e63971","Type":"ContainerStarted","Data":"acb527396cb316e35fa76097e8f1dbeda621bb250e35ce6a8055fd6164931b26"} Mar 11 13:42:48 crc kubenswrapper[4923]: I0311 13:42:48.832279 4923 generic.go:334] "Generic (PLEG): container finished" podID="b138d16e-9b69-4a49-8182-53bcb9346977" containerID="9ed051ae529bdae1271fb570d419074d3b72780bc9e75ed56af21badb86328c1" exitCode=1 Mar 11 13:42:48 crc kubenswrapper[4923]: I0311 13:42:48.832318 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"b138d16e-9b69-4a49-8182-53bcb9346977","Type":"ContainerDied","Data":"9ed051ae529bdae1271fb570d419074d3b72780bc9e75ed56af21badb86328c1"} Mar 11 13:42:48 crc kubenswrapper[4923]: I0311 13:42:48.832638 4923 scope.go:117] "RemoveContainer" containerID="c31b29598b06dcee0d8310749c3cf3bdb877be2865fba41550baadc936f9b3f8" Mar 11 13:42:48 crc kubenswrapper[4923]: I0311 13:42:48.832500 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="b138d16e-9b69-4a49-8182-53bcb9346977" containerName="cinder-api-log" containerID="cri-o://79905bbb8aebd892eadaf02ecc3305cf7393b8e1caf4211e0dbf6f70950da97e" gracePeriod=30 Mar 11 13:42:48 crc kubenswrapper[4923]: I0311 13:42:48.852272 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-scheduler-0" podStartSLOduration=2.852244356 podStartE2EDuration="2.852244356s" podCreationTimestamp="2026-03-11 13:42:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:42:48.844276 +0000 UTC m=+1216.526177724" watchObservedRunningTime="2026-03-11 13:42:48.852244356 +0000 UTC m=+1216.534146080" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.290411 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.486305 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b138d16e-9b69-4a49-8182-53bcb9346977-logs\") pod \"b138d16e-9b69-4a49-8182-53bcb9346977\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.486703 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7sgb\" (UniqueName: \"kubernetes.io/projected/b138d16e-9b69-4a49-8182-53bcb9346977-kube-api-access-c7sgb\") pod \"b138d16e-9b69-4a49-8182-53bcb9346977\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.486723 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b138d16e-9b69-4a49-8182-53bcb9346977-scripts\") pod \"b138d16e-9b69-4a49-8182-53bcb9346977\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.486939 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b138d16e-9b69-4a49-8182-53bcb9346977-logs" (OuterVolumeSpecName: "logs") pod "b138d16e-9b69-4a49-8182-53bcb9346977" (UID: "b138d16e-9b69-4a49-8182-53bcb9346977"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.487524 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b138d16e-9b69-4a49-8182-53bcb9346977-config-data-custom\") pod \"b138d16e-9b69-4a49-8182-53bcb9346977\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.487566 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b138d16e-9b69-4a49-8182-53bcb9346977-etc-machine-id\") pod \"b138d16e-9b69-4a49-8182-53bcb9346977\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.487600 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b138d16e-9b69-4a49-8182-53bcb9346977-config-data\") pod \"b138d16e-9b69-4a49-8182-53bcb9346977\" (UID: \"b138d16e-9b69-4a49-8182-53bcb9346977\") " Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.487765 4923 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b138d16e-9b69-4a49-8182-53bcb9346977-logs\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.487799 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b138d16e-9b69-4a49-8182-53bcb9346977-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b138d16e-9b69-4a49-8182-53bcb9346977" (UID: "b138d16e-9b69-4a49-8182-53bcb9346977"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.492854 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b138d16e-9b69-4a49-8182-53bcb9346977-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b138d16e-9b69-4a49-8182-53bcb9346977" (UID: "b138d16e-9b69-4a49-8182-53bcb9346977"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.493470 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b138d16e-9b69-4a49-8182-53bcb9346977-kube-api-access-c7sgb" (OuterVolumeSpecName: "kube-api-access-c7sgb") pod "b138d16e-9b69-4a49-8182-53bcb9346977" (UID: "b138d16e-9b69-4a49-8182-53bcb9346977"). InnerVolumeSpecName "kube-api-access-c7sgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.505282 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b138d16e-9b69-4a49-8182-53bcb9346977-scripts" (OuterVolumeSpecName: "scripts") pod "b138d16e-9b69-4a49-8182-53bcb9346977" (UID: "b138d16e-9b69-4a49-8182-53bcb9346977"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.537454 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b138d16e-9b69-4a49-8182-53bcb9346977-config-data" (OuterVolumeSpecName: "config-data") pod "b138d16e-9b69-4a49-8182-53bcb9346977" (UID: "b138d16e-9b69-4a49-8182-53bcb9346977"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.588183 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7sgb\" (UniqueName: \"kubernetes.io/projected/b138d16e-9b69-4a49-8182-53bcb9346977-kube-api-access-c7sgb\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.588213 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b138d16e-9b69-4a49-8182-53bcb9346977-scripts\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.588222 4923 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b138d16e-9b69-4a49-8182-53bcb9346977-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.588231 4923 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b138d16e-9b69-4a49-8182-53bcb9346977-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.588238 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b138d16e-9b69-4a49-8182-53bcb9346977-config-data\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.855712 4923 generic.go:334] "Generic (PLEG): container finished" podID="b138d16e-9b69-4a49-8182-53bcb9346977" containerID="79905bbb8aebd892eadaf02ecc3305cf7393b8e1caf4211e0dbf6f70950da97e" exitCode=143 Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.855844 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"b138d16e-9b69-4a49-8182-53bcb9346977","Type":"ContainerDied","Data":"79905bbb8aebd892eadaf02ecc3305cf7393b8e1caf4211e0dbf6f70950da97e"} Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.855897 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.856027 4923 scope.go:117] "RemoveContainer" containerID="9ed051ae529bdae1271fb570d419074d3b72780bc9e75ed56af21badb86328c1" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.855957 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"b138d16e-9b69-4a49-8182-53bcb9346977","Type":"ContainerDied","Data":"0eca812f9b38993489557623fe0e79a970b63d75630476385053c4bd6a3843e0"} Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.923973 4923 scope.go:117] "RemoveContainer" containerID="79905bbb8aebd892eadaf02ecc3305cf7393b8e1caf4211e0dbf6f70950da97e" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.961768 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.971657 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.973041 4923 scope.go:117] "RemoveContainer" containerID="9ed051ae529bdae1271fb570d419074d3b72780bc9e75ed56af21badb86328c1" Mar 11 13:42:49 crc kubenswrapper[4923]: E0311 13:42:49.976383 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ed051ae529bdae1271fb570d419074d3b72780bc9e75ed56af21badb86328c1\": container with ID starting with 9ed051ae529bdae1271fb570d419074d3b72780bc9e75ed56af21badb86328c1 not found: ID does not exist" containerID="9ed051ae529bdae1271fb570d419074d3b72780bc9e75ed56af21badb86328c1" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.976435 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ed051ae529bdae1271fb570d419074d3b72780bc9e75ed56af21badb86328c1"} err="failed to get container status \"9ed051ae529bdae1271fb570d419074d3b72780bc9e75ed56af21badb86328c1\": rpc error: code = NotFound desc = could not find container \"9ed051ae529bdae1271fb570d419074d3b72780bc9e75ed56af21badb86328c1\": container with ID starting with 9ed051ae529bdae1271fb570d419074d3b72780bc9e75ed56af21badb86328c1 not found: ID does not exist" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.976466 4923 scope.go:117] "RemoveContainer" containerID="79905bbb8aebd892eadaf02ecc3305cf7393b8e1caf4211e0dbf6f70950da97e" Mar 11 13:42:49 crc kubenswrapper[4923]: E0311 13:42:49.977053 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79905bbb8aebd892eadaf02ecc3305cf7393b8e1caf4211e0dbf6f70950da97e\": container with ID starting with 79905bbb8aebd892eadaf02ecc3305cf7393b8e1caf4211e0dbf6f70950da97e not found: ID does not exist" containerID="79905bbb8aebd892eadaf02ecc3305cf7393b8e1caf4211e0dbf6f70950da97e" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.977081 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79905bbb8aebd892eadaf02ecc3305cf7393b8e1caf4211e0dbf6f70950da97e"} err="failed to get container status \"79905bbb8aebd892eadaf02ecc3305cf7393b8e1caf4211e0dbf6f70950da97e\": rpc error: code = NotFound desc = could not find container \"79905bbb8aebd892eadaf02ecc3305cf7393b8e1caf4211e0dbf6f70950da97e\": container with ID starting with 79905bbb8aebd892eadaf02ecc3305cf7393b8e1caf4211e0dbf6f70950da97e not found: ID does not exist" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.985769 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Mar 11 13:42:49 crc kubenswrapper[4923]: E0311 13:42:49.986118 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b138d16e-9b69-4a49-8182-53bcb9346977" containerName="cinder-api" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.986133 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="b138d16e-9b69-4a49-8182-53bcb9346977" containerName="cinder-api" Mar 11 13:42:49 crc kubenswrapper[4923]: E0311 13:42:49.986149 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b138d16e-9b69-4a49-8182-53bcb9346977" containerName="cinder-api" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.986156 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="b138d16e-9b69-4a49-8182-53bcb9346977" containerName="cinder-api" Mar 11 13:42:49 crc kubenswrapper[4923]: E0311 13:42:49.986165 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b138d16e-9b69-4a49-8182-53bcb9346977" containerName="cinder-api-log" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.986171 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="b138d16e-9b69-4a49-8182-53bcb9346977" containerName="cinder-api-log" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.986307 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="b138d16e-9b69-4a49-8182-53bcb9346977" containerName="cinder-api" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.986320 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="b138d16e-9b69-4a49-8182-53bcb9346977" containerName="cinder-api" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.986331 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="b138d16e-9b69-4a49-8182-53bcb9346977" containerName="cinder-api-log" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.987121 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.993015 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"combined-ca-bundle" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.993045 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-api-config-data" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.993443 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.993073 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cert-cinder-public-svc" Mar 11 13:42:49 crc kubenswrapper[4923]: I0311 13:42:49.994176 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cert-cinder-internal-svc" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.094853 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-config-data\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.094912 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.094998 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-scripts\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.095132 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.095201 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-config-data-custom\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.095250 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2cf200ea-5fc3-4521-9d04-8662bc6783e7-logs\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.095283 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.095403 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqgrj\" (UniqueName: \"kubernetes.io/projected/2cf200ea-5fc3-4521-9d04-8662bc6783e7-kube-api-access-lqgrj\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.095452 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2cf200ea-5fc3-4521-9d04-8662bc6783e7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.196905 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-config-data\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.196944 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.197003 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-scripts\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.197057 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.197074 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-config-data-custom\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.197099 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2cf200ea-5fc3-4521-9d04-8662bc6783e7-logs\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.197113 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.197150 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqgrj\" (UniqueName: \"kubernetes.io/projected/2cf200ea-5fc3-4521-9d04-8662bc6783e7-kube-api-access-lqgrj\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.197167 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2cf200ea-5fc3-4521-9d04-8662bc6783e7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.197244 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2cf200ea-5fc3-4521-9d04-8662bc6783e7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.197773 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2cf200ea-5fc3-4521-9d04-8662bc6783e7-logs\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.201164 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.201408 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-config-data-custom\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.202512 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.202839 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.204749 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-config-data\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.209317 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-scripts\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.216793 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqgrj\" (UniqueName: \"kubernetes.io/projected/2cf200ea-5fc3-4521-9d04-8662bc6783e7-kube-api-access-lqgrj\") pod \"cinder-api-0\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.309925 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.795482 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Mar 11 13:42:50 crc kubenswrapper[4923]: I0311 13:42:50.868763 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"2cf200ea-5fc3-4521-9d04-8662bc6783e7","Type":"ContainerStarted","Data":"e646f158c5ba329d6986b91d6f562258f9bbd6e977c65ec4075c9fa22dd8a13b"} Mar 11 13:42:51 crc kubenswrapper[4923]: I0311 13:42:51.038531 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b138d16e-9b69-4a49-8182-53bcb9346977" path="/var/lib/kubelet/pods/b138d16e-9b69-4a49-8182-53bcb9346977/volumes" Mar 11 13:42:51 crc kubenswrapper[4923]: I0311 13:42:51.879493 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"2cf200ea-5fc3-4521-9d04-8662bc6783e7","Type":"ContainerStarted","Data":"704e6c4e48c5154d167ecabca60342e620e915fdadad03f457a6ba3f0d8efffd"} Mar 11 13:42:52 crc kubenswrapper[4923]: I0311 13:42:52.891977 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"2cf200ea-5fc3-4521-9d04-8662bc6783e7","Type":"ContainerStarted","Data":"93bfb560b4a2dffc02b367035772405d304824390a522eaf35a2dafee7c5ab69"} Mar 11 13:42:52 crc kubenswrapper[4923]: I0311 13:42:52.892370 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:42:52 crc kubenswrapper[4923]: I0311 13:42:52.930128 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-api-0" podStartSLOduration=3.930095987 podStartE2EDuration="3.930095987s" podCreationTimestamp="2026-03-11 13:42:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:42:52.926615227 +0000 UTC m=+1220.608516991" watchObservedRunningTime="2026-03-11 13:42:52.930095987 +0000 UTC m=+1220.611997751" Mar 11 13:42:54 crc kubenswrapper[4923]: I0311 13:42:54.128681 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:54 crc kubenswrapper[4923]: I0311 13:42:54.359900 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:54 crc kubenswrapper[4923]: I0311 13:42:54.976431 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Mar 11 13:42:55 crc kubenswrapper[4923]: I0311 13:42:55.928914 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-0" podUID="f87edc5e-c658-4ede-a5a2-157f3b97ee80" containerName="cinder-backup" containerID="cri-o://898fd024cb5a8479908125b72c3ba3ad34db8ecf31d0d64ed129377988f9982b" gracePeriod=30 Mar 11 13:42:55 crc kubenswrapper[4923]: I0311 13:42:55.928971 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-0" podUID="f87edc5e-c658-4ede-a5a2-157f3b97ee80" containerName="probe" containerID="cri-o://af06caaf9c4e3c325cb65dbb5745eb7e0c7abc06255512446c9595be987ac3c3" gracePeriod=30 Mar 11 13:42:56 crc kubenswrapper[4923]: I0311 13:42:56.940796 4923 generic.go:334] "Generic (PLEG): container finished" podID="f87edc5e-c658-4ede-a5a2-157f3b97ee80" containerID="af06caaf9c4e3c325cb65dbb5745eb7e0c7abc06255512446c9595be987ac3c3" exitCode=0 Mar 11 13:42:56 crc kubenswrapper[4923]: I0311 13:42:56.940869 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"f87edc5e-c658-4ede-a5a2-157f3b97ee80","Type":"ContainerDied","Data":"af06caaf9c4e3c325cb65dbb5745eb7e0c7abc06255512446c9595be987ac3c3"} Mar 11 13:42:57 crc kubenswrapper[4923]: I0311 13:42:57.022493 4923 scope.go:117] "RemoveContainer" containerID="8bf10a9ec7b294d63184d7c43bf1a0508fd8b54697b5139a5716aae8a4ba506c" Mar 11 13:42:57 crc kubenswrapper[4923]: I0311 13:42:57.022552 4923 scope.go:117] "RemoveContainer" containerID="4f323cc06c832423a935bd57211a21d27ae7dfdd630767d0c158db5de4fb1394" Mar 11 13:42:57 crc kubenswrapper[4923]: E0311 13:42:57.023042 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:42:57 crc kubenswrapper[4923]: I0311 13:42:57.225855 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:42:58 crc kubenswrapper[4923]: I0311 13:42:58.973665 4923 generic.go:334] "Generic (PLEG): container finished" podID="f87edc5e-c658-4ede-a5a2-157f3b97ee80" containerID="898fd024cb5a8479908125b72c3ba3ad34db8ecf31d0d64ed129377988f9982b" exitCode=0 Mar 11 13:42:58 crc kubenswrapper[4923]: I0311 13:42:58.973798 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"f87edc5e-c658-4ede-a5a2-157f3b97ee80","Type":"ContainerDied","Data":"898fd024cb5a8479908125b72c3ba3ad34db8ecf31d0d64ed129377988f9982b"} Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.075888 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.175305 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f87edc5e-c658-4ede-a5a2-157f3b97ee80-config-data\") pod \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.175367 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-etc-iscsi\") pod \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.175410 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-lib-modules\") pod \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.175474 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-sys\") pod \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.175600 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-dev\") pod \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.175629 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-var-locks-cinder\") pod \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.175495 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "f87edc5e-c658-4ede-a5a2-157f3b97ee80" (UID: "f87edc5e-c658-4ede-a5a2-157f3b97ee80"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.175686 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-dev" (OuterVolumeSpecName: "dev") pod "f87edc5e-c658-4ede-a5a2-157f3b97ee80" (UID: "f87edc5e-c658-4ede-a5a2-157f3b97ee80"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.175523 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "f87edc5e-c658-4ede-a5a2-157f3b97ee80" (UID: "f87edc5e-c658-4ede-a5a2-157f3b97ee80"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.175549 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-sys" (OuterVolumeSpecName: "sys") pod "f87edc5e-c658-4ede-a5a2-157f3b97ee80" (UID: "f87edc5e-c658-4ede-a5a2-157f3b97ee80"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.175739 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-etc-nvme\") pod \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.175749 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "f87edc5e-c658-4ede-a5a2-157f3b97ee80" (UID: "f87edc5e-c658-4ede-a5a2-157f3b97ee80"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.175811 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-var-locks-brick\") pod \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.175813 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "f87edc5e-c658-4ede-a5a2-157f3b97ee80" (UID: "f87edc5e-c658-4ede-a5a2-157f3b97ee80"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.175842 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kz2d\" (UniqueName: \"kubernetes.io/projected/f87edc5e-c658-4ede-a5a2-157f3b97ee80-kube-api-access-4kz2d\") pod \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.175857 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "f87edc5e-c658-4ede-a5a2-157f3b97ee80" (UID: "f87edc5e-c658-4ede-a5a2-157f3b97ee80"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.175893 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-run\") pod \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.176099 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-run" (OuterVolumeSpecName: "run") pod "f87edc5e-c658-4ede-a5a2-157f3b97ee80" (UID: "f87edc5e-c658-4ede-a5a2-157f3b97ee80"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.176532 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f87edc5e-c658-4ede-a5a2-157f3b97ee80-scripts\") pod \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.176590 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f87edc5e-c658-4ede-a5a2-157f3b97ee80-config-data-custom\") pod \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.176635 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-var-lib-cinder\") pod \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.176661 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-etc-machine-id\") pod \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\" (UID: \"f87edc5e-c658-4ede-a5a2-157f3b97ee80\") " Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.177137 4923 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-etc-iscsi\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.177161 4923 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-lib-modules\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.177175 4923 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-sys\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.177218 4923 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-dev\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.177234 4923 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.177250 4923 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-etc-nvme\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.177265 4923 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-var-locks-brick\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.177279 4923 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-run\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.177508 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f87edc5e-c658-4ede-a5a2-157f3b97ee80" (UID: "f87edc5e-c658-4ede-a5a2-157f3b97ee80"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.178120 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "f87edc5e-c658-4ede-a5a2-157f3b97ee80" (UID: "f87edc5e-c658-4ede-a5a2-157f3b97ee80"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.183174 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f87edc5e-c658-4ede-a5a2-157f3b97ee80-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f87edc5e-c658-4ede-a5a2-157f3b97ee80" (UID: "f87edc5e-c658-4ede-a5a2-157f3b97ee80"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.183748 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f87edc5e-c658-4ede-a5a2-157f3b97ee80-kube-api-access-4kz2d" (OuterVolumeSpecName: "kube-api-access-4kz2d") pod "f87edc5e-c658-4ede-a5a2-157f3b97ee80" (UID: "f87edc5e-c658-4ede-a5a2-157f3b97ee80"). InnerVolumeSpecName "kube-api-access-4kz2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.185470 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f87edc5e-c658-4ede-a5a2-157f3b97ee80-scripts" (OuterVolumeSpecName: "scripts") pod "f87edc5e-c658-4ede-a5a2-157f3b97ee80" (UID: "f87edc5e-c658-4ede-a5a2-157f3b97ee80"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.254250 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f87edc5e-c658-4ede-a5a2-157f3b97ee80-config-data" (OuterVolumeSpecName: "config-data") pod "f87edc5e-c658-4ede-a5a2-157f3b97ee80" (UID: "f87edc5e-c658-4ede-a5a2-157f3b97ee80"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.278958 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kz2d\" (UniqueName: \"kubernetes.io/projected/f87edc5e-c658-4ede-a5a2-157f3b97ee80-kube-api-access-4kz2d\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.279009 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f87edc5e-c658-4ede-a5a2-157f3b97ee80-scripts\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.279026 4923 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f87edc5e-c658-4ede-a5a2-157f3b97ee80-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.279040 4923 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.279050 4923 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f87edc5e-c658-4ede-a5a2-157f3b97ee80-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.279060 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f87edc5e-c658-4ede-a5a2-157f3b97ee80-config-data\") on node \"crc\" DevicePath \"\"" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.988911 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"f87edc5e-c658-4ede-a5a2-157f3b97ee80","Type":"ContainerDied","Data":"73e2bba53e451c81665f0b82c801fe1d9becc3d3bba38f4971178daa37bb2706"} Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.988963 4923 scope.go:117] "RemoveContainer" containerID="af06caaf9c4e3c325cb65dbb5745eb7e0c7abc06255512446c9595be987ac3c3" Mar 11 13:42:59 crc kubenswrapper[4923]: I0311 13:42:59.989022 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.035813 4923 scope.go:117] "RemoveContainer" containerID="898fd024cb5a8479908125b72c3ba3ad34db8ecf31d0d64ed129377988f9982b" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.052587 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.065768 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.083439 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Mar 11 13:43:00 crc kubenswrapper[4923]: E0311 13:43:00.083964 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f87edc5e-c658-4ede-a5a2-157f3b97ee80" containerName="probe" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.084009 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f87edc5e-c658-4ede-a5a2-157f3b97ee80" containerName="probe" Mar 11 13:43:00 crc kubenswrapper[4923]: E0311 13:43:00.084035 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f87edc5e-c658-4ede-a5a2-157f3b97ee80" containerName="cinder-backup" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.084052 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f87edc5e-c658-4ede-a5a2-157f3b97ee80" containerName="cinder-backup" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.084430 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f87edc5e-c658-4ede-a5a2-157f3b97ee80" containerName="probe" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.084491 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f87edc5e-c658-4ede-a5a2-157f3b97ee80" containerName="cinder-backup" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.086143 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.088721 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-backup-config-data" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.106251 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.192389 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-scripts\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.192429 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-config-data\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.192449 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkgx8\" (UniqueName: \"kubernetes.io/projected/722b7fb9-8dea-4dfb-9d69-d73216d237a6-kube-api-access-vkgx8\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.192481 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-lib-modules\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.192497 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-run\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.192512 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.192532 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.192743 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-dev\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.192895 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.192950 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-etc-nvme\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.193004 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-sys\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.193073 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-config-data-custom\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.193173 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.193240 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.193335 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.294529 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.294586 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.294625 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.294688 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-scripts\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.294705 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.294716 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-config-data\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.294803 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkgx8\" (UniqueName: \"kubernetes.io/projected/722b7fb9-8dea-4dfb-9d69-d73216d237a6-kube-api-access-vkgx8\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.294881 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-lib-modules\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.294873 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.294920 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-run\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.294963 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.295019 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.295092 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-dev\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.295180 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.295211 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-etc-nvme\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.295243 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-sys\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.295288 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-config-data-custom\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.295527 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.295940 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-lib-modules\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.295996 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-run\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.296065 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.296130 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.296182 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-dev\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.297085 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-etc-nvme\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.297186 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-sys\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.301583 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-config-data-custom\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.302453 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-config-data\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.305137 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-scripts\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.311983 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.321133 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkgx8\" (UniqueName: \"kubernetes.io/projected/722b7fb9-8dea-4dfb-9d69-d73216d237a6-kube-api-access-vkgx8\") pod \"cinder-backup-0\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.450775 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.959557 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Mar 11 13:43:00 crc kubenswrapper[4923]: W0311 13:43:00.963638 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod722b7fb9_8dea_4dfb_9d69_d73216d237a6.slice/crio-499ebd2938f5d599c4a8f1fe52435007565ace1bbd40b08c619fe9e4a3f98b6c WatchSource:0}: Error finding container 499ebd2938f5d599c4a8f1fe52435007565ace1bbd40b08c619fe9e4a3f98b6c: Status 404 returned error can't find the container with id 499ebd2938f5d599c4a8f1fe52435007565ace1bbd40b08c619fe9e4a3f98b6c Mar 11 13:43:00 crc kubenswrapper[4923]: I0311 13:43:00.999768 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"722b7fb9-8dea-4dfb-9d69-d73216d237a6","Type":"ContainerStarted","Data":"499ebd2938f5d599c4a8f1fe52435007565ace1bbd40b08c619fe9e4a3f98b6c"} Mar 11 13:43:01 crc kubenswrapper[4923]: I0311 13:43:01.035102 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f87edc5e-c658-4ede-a5a2-157f3b97ee80" path="/var/lib/kubelet/pods/f87edc5e-c658-4ede-a5a2-157f3b97ee80/volumes" Mar 11 13:43:02 crc kubenswrapper[4923]: I0311 13:43:02.028096 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"722b7fb9-8dea-4dfb-9d69-d73216d237a6","Type":"ContainerStarted","Data":"b6c5c5484239241499cb9d89d3783a167a592b3af17fa67e8c775ae590ce7c43"} Mar 11 13:43:02 crc kubenswrapper[4923]: I0311 13:43:02.028721 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"722b7fb9-8dea-4dfb-9d69-d73216d237a6","Type":"ContainerStarted","Data":"2986755d10efa4502902211ac0a8fd41a81ad8db127f9a415998952f7619cfc1"} Mar 11 13:43:02 crc kubenswrapper[4923]: I0311 13:43:02.065140 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-backup-0" podStartSLOduration=2.065113819 podStartE2EDuration="2.065113819s" podCreationTimestamp="2026-03-11 13:43:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:43:02.052942635 +0000 UTC m=+1229.734844389" watchObservedRunningTime="2026-03-11 13:43:02.065113819 +0000 UTC m=+1229.747015543" Mar 11 13:43:07 crc kubenswrapper[4923]: I0311 13:43:07.436110 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:07 crc kubenswrapper[4923]: I0311 13:43:07.509933 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Mar 11 13:43:08 crc kubenswrapper[4923]: I0311 13:43:08.077994 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-0" podUID="97b85261-6521-4ca2-b7e9-b6fc59e63971" containerName="cinder-scheduler" containerID="cri-o://acb527396cb316e35fa76097e8f1dbeda621bb250e35ce6a8055fd6164931b26" gracePeriod=30 Mar 11 13:43:08 crc kubenswrapper[4923]: I0311 13:43:08.078033 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-0" podUID="97b85261-6521-4ca2-b7e9-b6fc59e63971" containerName="probe" containerID="cri-o://4824ba162fa478a80e5086b293c197316a3e376385333aa14602521ad8e30f25" gracePeriod=30 Mar 11 13:43:09 crc kubenswrapper[4923]: I0311 13:43:09.088983 4923 generic.go:334] "Generic (PLEG): container finished" podID="97b85261-6521-4ca2-b7e9-b6fc59e63971" containerID="4824ba162fa478a80e5086b293c197316a3e376385333aa14602521ad8e30f25" exitCode=0 Mar 11 13:43:09 crc kubenswrapper[4923]: I0311 13:43:09.089285 4923 generic.go:334] "Generic (PLEG): container finished" podID="97b85261-6521-4ca2-b7e9-b6fc59e63971" containerID="acb527396cb316e35fa76097e8f1dbeda621bb250e35ce6a8055fd6164931b26" exitCode=0 Mar 11 13:43:09 crc kubenswrapper[4923]: I0311 13:43:09.089060 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"97b85261-6521-4ca2-b7e9-b6fc59e63971","Type":"ContainerDied","Data":"4824ba162fa478a80e5086b293c197316a3e376385333aa14602521ad8e30f25"} Mar 11 13:43:09 crc kubenswrapper[4923]: I0311 13:43:09.089316 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"97b85261-6521-4ca2-b7e9-b6fc59e63971","Type":"ContainerDied","Data":"acb527396cb316e35fa76097e8f1dbeda621bb250e35ce6a8055fd6164931b26"} Mar 11 13:43:09 crc kubenswrapper[4923]: I0311 13:43:09.089326 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"97b85261-6521-4ca2-b7e9-b6fc59e63971","Type":"ContainerDied","Data":"3b88a725c03ea4013d509461c020d6da59240b457be505dd9003478a180f27b8"} Mar 11 13:43:09 crc kubenswrapper[4923]: I0311 13:43:09.089338 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b88a725c03ea4013d509461c020d6da59240b457be505dd9003478a180f27b8" Mar 11 13:43:09 crc kubenswrapper[4923]: I0311 13:43:09.111425 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:09 crc kubenswrapper[4923]: I0311 13:43:09.259019 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/97b85261-6521-4ca2-b7e9-b6fc59e63971-etc-machine-id\") pod \"97b85261-6521-4ca2-b7e9-b6fc59e63971\" (UID: \"97b85261-6521-4ca2-b7e9-b6fc59e63971\") " Mar 11 13:43:09 crc kubenswrapper[4923]: I0311 13:43:09.259167 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h64tg\" (UniqueName: \"kubernetes.io/projected/97b85261-6521-4ca2-b7e9-b6fc59e63971-kube-api-access-h64tg\") pod \"97b85261-6521-4ca2-b7e9-b6fc59e63971\" (UID: \"97b85261-6521-4ca2-b7e9-b6fc59e63971\") " Mar 11 13:43:09 crc kubenswrapper[4923]: I0311 13:43:09.259212 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97b85261-6521-4ca2-b7e9-b6fc59e63971-config-data\") pod \"97b85261-6521-4ca2-b7e9-b6fc59e63971\" (UID: \"97b85261-6521-4ca2-b7e9-b6fc59e63971\") " Mar 11 13:43:09 crc kubenswrapper[4923]: I0311 13:43:09.259466 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/97b85261-6521-4ca2-b7e9-b6fc59e63971-config-data-custom\") pod \"97b85261-6521-4ca2-b7e9-b6fc59e63971\" (UID: \"97b85261-6521-4ca2-b7e9-b6fc59e63971\") " Mar 11 13:43:09 crc kubenswrapper[4923]: I0311 13:43:09.259512 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97b85261-6521-4ca2-b7e9-b6fc59e63971-scripts\") pod \"97b85261-6521-4ca2-b7e9-b6fc59e63971\" (UID: \"97b85261-6521-4ca2-b7e9-b6fc59e63971\") " Mar 11 13:43:09 crc kubenswrapper[4923]: I0311 13:43:09.260565 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97b85261-6521-4ca2-b7e9-b6fc59e63971-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "97b85261-6521-4ca2-b7e9-b6fc59e63971" (UID: "97b85261-6521-4ca2-b7e9-b6fc59e63971"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:09 crc kubenswrapper[4923]: I0311 13:43:09.275123 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97b85261-6521-4ca2-b7e9-b6fc59e63971-scripts" (OuterVolumeSpecName: "scripts") pod "97b85261-6521-4ca2-b7e9-b6fc59e63971" (UID: "97b85261-6521-4ca2-b7e9-b6fc59e63971"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:43:09 crc kubenswrapper[4923]: I0311 13:43:09.275190 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97b85261-6521-4ca2-b7e9-b6fc59e63971-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "97b85261-6521-4ca2-b7e9-b6fc59e63971" (UID: "97b85261-6521-4ca2-b7e9-b6fc59e63971"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:43:09 crc kubenswrapper[4923]: I0311 13:43:09.275141 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97b85261-6521-4ca2-b7e9-b6fc59e63971-kube-api-access-h64tg" (OuterVolumeSpecName: "kube-api-access-h64tg") pod "97b85261-6521-4ca2-b7e9-b6fc59e63971" (UID: "97b85261-6521-4ca2-b7e9-b6fc59e63971"). InnerVolumeSpecName "kube-api-access-h64tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:43:09 crc kubenswrapper[4923]: I0311 13:43:09.334448 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97b85261-6521-4ca2-b7e9-b6fc59e63971-config-data" (OuterVolumeSpecName: "config-data") pod "97b85261-6521-4ca2-b7e9-b6fc59e63971" (UID: "97b85261-6521-4ca2-b7e9-b6fc59e63971"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:43:09 crc kubenswrapper[4923]: I0311 13:43:09.360918 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h64tg\" (UniqueName: \"kubernetes.io/projected/97b85261-6521-4ca2-b7e9-b6fc59e63971-kube-api-access-h64tg\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:09 crc kubenswrapper[4923]: I0311 13:43:09.360959 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97b85261-6521-4ca2-b7e9-b6fc59e63971-config-data\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:09 crc kubenswrapper[4923]: I0311 13:43:09.360972 4923 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/97b85261-6521-4ca2-b7e9-b6fc59e63971-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:09 crc kubenswrapper[4923]: I0311 13:43:09.360982 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97b85261-6521-4ca2-b7e9-b6fc59e63971-scripts\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:09 crc kubenswrapper[4923]: I0311 13:43:09.360992 4923 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/97b85261-6521-4ca2-b7e9-b6fc59e63971-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.097817 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.153937 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.160850 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.186043 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Mar 11 13:43:10 crc kubenswrapper[4923]: E0311 13:43:10.186439 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97b85261-6521-4ca2-b7e9-b6fc59e63971" containerName="probe" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.186492 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="97b85261-6521-4ca2-b7e9-b6fc59e63971" containerName="probe" Mar 11 13:43:10 crc kubenswrapper[4923]: E0311 13:43:10.186511 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97b85261-6521-4ca2-b7e9-b6fc59e63971" containerName="cinder-scheduler" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.186521 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="97b85261-6521-4ca2-b7e9-b6fc59e63971" containerName="cinder-scheduler" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.186668 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="97b85261-6521-4ca2-b7e9-b6fc59e63971" containerName="cinder-scheduler" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.186684 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="97b85261-6521-4ca2-b7e9-b6fc59e63971" containerName="probe" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.187605 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.189589 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cinder-kuttl-tests"/"cinder-scheduler-config-data" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.210858 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.279083 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-scripts\") pod \"cinder-scheduler-0\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.279158 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.279201 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zqvf\" (UniqueName: \"kubernetes.io/projected/12c7d96d-0e1a-4059-9e54-2a8438917bad-kube-api-access-9zqvf\") pod \"cinder-scheduler-0\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.279513 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-config-data\") pod \"cinder-scheduler-0\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.279661 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12c7d96d-0e1a-4059-9e54-2a8438917bad-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.279749 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.381085 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12c7d96d-0e1a-4059-9e54-2a8438917bad-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.381153 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.381180 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-scripts\") pod \"cinder-scheduler-0\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.381215 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.381206 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12c7d96d-0e1a-4059-9e54-2a8438917bad-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.381250 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zqvf\" (UniqueName: \"kubernetes.io/projected/12c7d96d-0e1a-4059-9e54-2a8438917bad-kube-api-access-9zqvf\") pod \"cinder-scheduler-0\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.381572 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-config-data\") pod \"cinder-scheduler-0\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.386765 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.386967 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-scripts\") pod \"cinder-scheduler-0\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.387169 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-config-data\") pod \"cinder-scheduler-0\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.394934 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.415646 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zqvf\" (UniqueName: \"kubernetes.io/projected/12c7d96d-0e1a-4059-9e54-2a8438917bad-kube-api-access-9zqvf\") pod \"cinder-scheduler-0\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.451909 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:10 crc kubenswrapper[4923]: I0311 13:43:10.516749 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:11 crc kubenswrapper[4923]: I0311 13:43:11.041911 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97b85261-6521-4ca2-b7e9-b6fc59e63971" path="/var/lib/kubelet/pods/97b85261-6521-4ca2-b7e9-b6fc59e63971/volumes" Mar 11 13:43:11 crc kubenswrapper[4923]: I0311 13:43:11.582758 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Mar 11 13:43:12 crc kubenswrapper[4923]: I0311 13:43:12.022558 4923 scope.go:117] "RemoveContainer" containerID="8bf10a9ec7b294d63184d7c43bf1a0508fd8b54697b5139a5716aae8a4ba506c" Mar 11 13:43:12 crc kubenswrapper[4923]: I0311 13:43:12.022590 4923 scope.go:117] "RemoveContainer" containerID="4f323cc06c832423a935bd57211a21d27ae7dfdd630767d0c158db5de4fb1394" Mar 11 13:43:12 crc kubenswrapper[4923]: E0311 13:43:12.022930 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:43:12 crc kubenswrapper[4923]: I0311 13:43:12.119906 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"12c7d96d-0e1a-4059-9e54-2a8438917bad","Type":"ContainerStarted","Data":"58a54cd9a9104dcb279c12e8420fff8bf4b9f16614f8d75171a520ffa931f350"} Mar 11 13:43:12 crc kubenswrapper[4923]: I0311 13:43:12.120303 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"12c7d96d-0e1a-4059-9e54-2a8438917bad","Type":"ContainerStarted","Data":"161c4c9a7f81dd4ee3a442d19a38488aba85b8beba4e404eb1f6d36c76b3c73a"} Mar 11 13:43:12 crc kubenswrapper[4923]: I0311 13:43:12.152138 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:43:13 crc kubenswrapper[4923]: I0311 13:43:13.133599 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"12c7d96d-0e1a-4059-9e54-2a8438917bad","Type":"ContainerStarted","Data":"f6c9cd39e8280d1ff398def00133a920696fff51cced1383a3a022c0bd2d2284"} Mar 11 13:43:13 crc kubenswrapper[4923]: I0311 13:43:13.159339 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cinder-scheduler-0" podStartSLOduration=3.159308236 podStartE2EDuration="3.159308236s" podCreationTimestamp="2026-03-11 13:43:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:43:13.156842541 +0000 UTC m=+1240.838744265" watchObservedRunningTime="2026-03-11 13:43:13.159308236 +0000 UTC m=+1240.841209990" Mar 11 13:43:20 crc kubenswrapper[4923]: I0311 13:43:20.517716 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:20 crc kubenswrapper[4923]: I0311 13:43:20.672031 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:26 crc kubenswrapper[4923]: I0311 13:43:26.022061 4923 scope.go:117] "RemoveContainer" containerID="8bf10a9ec7b294d63184d7c43bf1a0508fd8b54697b5139a5716aae8a4ba506c" Mar 11 13:43:26 crc kubenswrapper[4923]: I0311 13:43:26.022646 4923 scope.go:117] "RemoveContainer" containerID="4f323cc06c832423a935bd57211a21d27ae7dfdd630767d0c158db5de4fb1394" Mar 11 13:43:26 crc kubenswrapper[4923]: E0311 13:43:26.022991 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"cinder-volume\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=cinder-volume pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\", failed to \"StartContainer\" for \"probe\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=probe pod=cinder-volume-volume1-0_cinder-kuttl-tests(ed2ffa6a-46e2-4257-a8c0-54d2808b37c4)\"]" pod="cinder-kuttl-tests/cinder-volume-volume1-0" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" Mar 11 13:43:30 crc kubenswrapper[4923]: I0311 13:43:30.790014 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.078732 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-bh8nh"] Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.085059 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-db-sync-bh8nh"] Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.129592 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.129978 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-0" podUID="12c7d96d-0e1a-4059-9e54-2a8438917bad" containerName="cinder-scheduler" containerID="cri-o://58a54cd9a9104dcb279c12e8420fff8bf4b9f16614f8d75171a520ffa931f350" gracePeriod=30 Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.130158 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-scheduler-0" podUID="12c7d96d-0e1a-4059-9e54-2a8438917bad" containerName="probe" containerID="cri-o://f6c9cd39e8280d1ff398def00133a920696fff51cced1383a3a022c0bd2d2284" gracePeriod=30 Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.165557 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cinder-kuttl-tests/cindere6fe-account-delete-57xck"] Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.166699 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cindere6fe-account-delete-57xck" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.179896 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.180322 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-0" podUID="722b7fb9-8dea-4dfb-9d69-d73216d237a6" containerName="cinder-backup" containerID="cri-o://2986755d10efa4502902211ac0a8fd41a81ad8db127f9a415998952f7619cfc1" gracePeriod=30 Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.180542 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-backup-0" podUID="722b7fb9-8dea-4dfb-9d69-d73216d237a6" containerName="probe" containerID="cri-o://b6c5c5484239241499cb9d89d3783a167a592b3af17fa67e8c775ae590ce7c43" gracePeriod=30 Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.189812 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.201414 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.202242 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="2cf200ea-5fc3-4521-9d04-8662bc6783e7" containerName="cinder-api-log" containerID="cri-o://704e6c4e48c5154d167ecabca60342e620e915fdadad03f457a6ba3f0d8efffd" gracePeriod=30 Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.202493 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="cinder-kuttl-tests/cinder-api-0" podUID="2cf200ea-5fc3-4521-9d04-8662bc6783e7" containerName="cinder-api" containerID="cri-o://93bfb560b4a2dffc02b367035772405d304824390a522eaf35a2dafee7c5ab69" gracePeriod=30 Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.208470 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cindere6fe-account-delete-57xck"] Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.264139 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d764fb3-601e-4159-af7b-75c833e2792f-operator-scripts\") pod \"cindere6fe-account-delete-57xck\" (UID: \"9d764fb3-601e-4159-af7b-75c833e2792f\") " pod="cinder-kuttl-tests/cindere6fe-account-delete-57xck" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.264500 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9p8r\" (UniqueName: \"kubernetes.io/projected/9d764fb3-601e-4159-af7b-75c833e2792f-kube-api-access-n9p8r\") pod \"cindere6fe-account-delete-57xck\" (UID: \"9d764fb3-601e-4159-af7b-75c833e2792f\") " pod="cinder-kuttl-tests/cindere6fe-account-delete-57xck" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.337761 4923 generic.go:334] "Generic (PLEG): container finished" podID="2cf200ea-5fc3-4521-9d04-8662bc6783e7" containerID="704e6c4e48c5154d167ecabca60342e620e915fdadad03f457a6ba3f0d8efffd" exitCode=143 Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.337903 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"2cf200ea-5fc3-4521-9d04-8662bc6783e7","Type":"ContainerDied","Data":"704e6c4e48c5154d167ecabca60342e620e915fdadad03f457a6ba3f0d8efffd"} Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.366733 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d764fb3-601e-4159-af7b-75c833e2792f-operator-scripts\") pod \"cindere6fe-account-delete-57xck\" (UID: \"9d764fb3-601e-4159-af7b-75c833e2792f\") " pod="cinder-kuttl-tests/cindere6fe-account-delete-57xck" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.366828 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9p8r\" (UniqueName: \"kubernetes.io/projected/9d764fb3-601e-4159-af7b-75c833e2792f-kube-api-access-n9p8r\") pod \"cindere6fe-account-delete-57xck\" (UID: \"9d764fb3-601e-4159-af7b-75c833e2792f\") " pod="cinder-kuttl-tests/cindere6fe-account-delete-57xck" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.367783 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d764fb3-601e-4159-af7b-75c833e2792f-operator-scripts\") pod \"cindere6fe-account-delete-57xck\" (UID: \"9d764fb3-601e-4159-af7b-75c833e2792f\") " pod="cinder-kuttl-tests/cindere6fe-account-delete-57xck" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.395067 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9p8r\" (UniqueName: \"kubernetes.io/projected/9d764fb3-601e-4159-af7b-75c833e2792f-kube-api-access-n9p8r\") pod \"cindere6fe-account-delete-57xck\" (UID: \"9d764fb3-601e-4159-af7b-75c833e2792f\") " pod="cinder-kuttl-tests/cindere6fe-account-delete-57xck" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.496132 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cindere6fe-account-delete-57xck" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.604938 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.672385 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-var-lib-cinder\") pod \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.672760 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-etc-iscsi\") pod \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.672801 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-var-locks-cinder\") pod \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.672872 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-run\") pod \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.672872 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" (UID: "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.672896 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-var-locks-brick\") pod \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.672915 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-run" (OuterVolumeSpecName: "run") pod "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" (UID: "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.672920 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" (UID: "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.672934 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7hfj\" (UniqueName: \"kubernetes.io/projected/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-kube-api-access-h7hfj\") pod \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.672948 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" (UID: "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.672956 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-etc-nvme\") pod \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.672981 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-dev\") pod \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.673002 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-sys\") pod \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.673005 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" (UID: "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.673028 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-dev" (OuterVolumeSpecName: "dev") pod "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" (UID: "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.673037 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-lib-modules\") pod \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.673045 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-sys" (OuterVolumeSpecName: "sys") pod "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" (UID: "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.673079 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-scripts\") pod \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.673132 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" (UID: "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.673141 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-etc-machine-id\") pod \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.673170 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-config-data-custom\") pod \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.673209 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-config-data\") pod \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\" (UID: \"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4\") " Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.673527 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" (UID: "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.673759 4923 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-sys\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.673780 4923 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-lib-modules\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.673789 4923 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.673797 4923 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-etc-iscsi\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.673805 4923 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.673814 4923 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-run\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.673821 4923 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-var-locks-brick\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.673831 4923 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-etc-nvme\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.673838 4923 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-dev\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.675514 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" (UID: "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.676054 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-kube-api-access-h7hfj" (OuterVolumeSpecName: "kube-api-access-h7hfj") pod "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" (UID: "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4"). InnerVolumeSpecName "kube-api-access-h7hfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.676472 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-scripts" (OuterVolumeSpecName: "scripts") pod "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" (UID: "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.676648 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" (UID: "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.755269 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-config-data" (OuterVolumeSpecName: "config-data") pod "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" (UID: "ed2ffa6a-46e2-4257-a8c0-54d2808b37c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.775620 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7hfj\" (UniqueName: \"kubernetes.io/projected/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-kube-api-access-h7hfj\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.775650 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-scripts\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.775659 4923 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.775667 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-config-data\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.775676 4923 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:32 crc kubenswrapper[4923]: I0311 13:43:32.970812 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cinder-kuttl-tests/cindere6fe-account-delete-57xck"] Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.035291 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86738f89-b8e9-4cd5-806a-a54529ef1078" path="/var/lib/kubelet/pods/86738f89-b8e9-4cd5-806a-a54529ef1078/volumes" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.174518 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.282637 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-run\") pod \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.282726 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-scripts\") pod \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.282754 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-var-lib-cinder\") pod \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.282792 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-etc-iscsi\") pod \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.282830 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-combined-ca-bundle\") pod \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.282857 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-lib-modules\") pod \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.282906 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-var-locks-cinder\") pod \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.282956 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-var-locks-brick\") pod \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.283002 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-etc-machine-id\") pod \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.283039 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-config-data-custom\") pod \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.283072 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-config-data\") pod \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.283100 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkgx8\" (UniqueName: \"kubernetes.io/projected/722b7fb9-8dea-4dfb-9d69-d73216d237a6-kube-api-access-vkgx8\") pod \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.283077 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "722b7fb9-8dea-4dfb-9d69-d73216d237a6" (UID: "722b7fb9-8dea-4dfb-9d69-d73216d237a6"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.283147 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-etc-nvme\") pod \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.283167 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-sys\") pod \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.283166 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-var-lib-cinder" (OuterVolumeSpecName: "var-lib-cinder") pod "722b7fb9-8dea-4dfb-9d69-d73216d237a6" (UID: "722b7fb9-8dea-4dfb-9d69-d73216d237a6"). InnerVolumeSpecName "var-lib-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.283202 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "722b7fb9-8dea-4dfb-9d69-d73216d237a6" (UID: "722b7fb9-8dea-4dfb-9d69-d73216d237a6"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.283211 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-dev\") pod \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\" (UID: \"722b7fb9-8dea-4dfb-9d69-d73216d237a6\") " Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.283589 4923 reconciler_common.go:293] "Volume detached for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-var-lib-cinder\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.283608 4923 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-etc-iscsi\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.283620 4923 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-lib-modules\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.283672 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-dev" (OuterVolumeSpecName: "dev") pod "722b7fb9-8dea-4dfb-9d69-d73216d237a6" (UID: "722b7fb9-8dea-4dfb-9d69-d73216d237a6"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.283719 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "722b7fb9-8dea-4dfb-9d69-d73216d237a6" (UID: "722b7fb9-8dea-4dfb-9d69-d73216d237a6"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.283735 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-run" (OuterVolumeSpecName: "run") pod "722b7fb9-8dea-4dfb-9d69-d73216d237a6" (UID: "722b7fb9-8dea-4dfb-9d69-d73216d237a6"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.283769 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "722b7fb9-8dea-4dfb-9d69-d73216d237a6" (UID: "722b7fb9-8dea-4dfb-9d69-d73216d237a6"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.283746 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-sys" (OuterVolumeSpecName: "sys") pod "722b7fb9-8dea-4dfb-9d69-d73216d237a6" (UID: "722b7fb9-8dea-4dfb-9d69-d73216d237a6"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.283791 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-var-locks-cinder" (OuterVolumeSpecName: "var-locks-cinder") pod "722b7fb9-8dea-4dfb-9d69-d73216d237a6" (UID: "722b7fb9-8dea-4dfb-9d69-d73216d237a6"). InnerVolumeSpecName "var-locks-cinder". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.283816 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "722b7fb9-8dea-4dfb-9d69-d73216d237a6" (UID: "722b7fb9-8dea-4dfb-9d69-d73216d237a6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.289083 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-scripts" (OuterVolumeSpecName: "scripts") pod "722b7fb9-8dea-4dfb-9d69-d73216d237a6" (UID: "722b7fb9-8dea-4dfb-9d69-d73216d237a6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.289639 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/722b7fb9-8dea-4dfb-9d69-d73216d237a6-kube-api-access-vkgx8" (OuterVolumeSpecName: "kube-api-access-vkgx8") pod "722b7fb9-8dea-4dfb-9d69-d73216d237a6" (UID: "722b7fb9-8dea-4dfb-9d69-d73216d237a6"). InnerVolumeSpecName "kube-api-access-vkgx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.295043 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "722b7fb9-8dea-4dfb-9d69-d73216d237a6" (UID: "722b7fb9-8dea-4dfb-9d69-d73216d237a6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.325171 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "722b7fb9-8dea-4dfb-9d69-d73216d237a6" (UID: "722b7fb9-8dea-4dfb-9d69-d73216d237a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.348973 4923 generic.go:334] "Generic (PLEG): container finished" podID="722b7fb9-8dea-4dfb-9d69-d73216d237a6" containerID="b6c5c5484239241499cb9d89d3783a167a592b3af17fa67e8c775ae590ce7c43" exitCode=0 Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.349005 4923 generic.go:334] "Generic (PLEG): container finished" podID="722b7fb9-8dea-4dfb-9d69-d73216d237a6" containerID="2986755d10efa4502902211ac0a8fd41a81ad8db127f9a415998952f7619cfc1" exitCode=0 Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.349050 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"722b7fb9-8dea-4dfb-9d69-d73216d237a6","Type":"ContainerDied","Data":"b6c5c5484239241499cb9d89d3783a167a592b3af17fa67e8c775ae590ce7c43"} Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.349077 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"722b7fb9-8dea-4dfb-9d69-d73216d237a6","Type":"ContainerDied","Data":"2986755d10efa4502902211ac0a8fd41a81ad8db127f9a415998952f7619cfc1"} Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.349087 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-backup-0" event={"ID":"722b7fb9-8dea-4dfb-9d69-d73216d237a6","Type":"ContainerDied","Data":"499ebd2938f5d599c4a8f1fe52435007565ace1bbd40b08c619fe9e4a3f98b6c"} Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.349105 4923 scope.go:117] "RemoveContainer" containerID="b6c5c5484239241499cb9d89d3783a167a592b3af17fa67e8c775ae590ce7c43" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.349293 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-backup-0" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.359152 4923 generic.go:334] "Generic (PLEG): container finished" podID="12c7d96d-0e1a-4059-9e54-2a8438917bad" containerID="f6c9cd39e8280d1ff398def00133a920696fff51cced1383a3a022c0bd2d2284" exitCode=0 Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.359392 4923 generic.go:334] "Generic (PLEG): container finished" podID="12c7d96d-0e1a-4059-9e54-2a8438917bad" containerID="58a54cd9a9104dcb279c12e8420fff8bf4b9f16614f8d75171a520ffa931f350" exitCode=0 Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.359233 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"12c7d96d-0e1a-4059-9e54-2a8438917bad","Type":"ContainerDied","Data":"f6c9cd39e8280d1ff398def00133a920696fff51cced1383a3a022c0bd2d2284"} Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.359497 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"12c7d96d-0e1a-4059-9e54-2a8438917bad","Type":"ContainerDied","Data":"58a54cd9a9104dcb279c12e8420fff8bf4b9f16614f8d75171a520ffa931f350"} Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.361559 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-volume-volume1-0" event={"ID":"ed2ffa6a-46e2-4257-a8c0-54d2808b37c4","Type":"ContainerDied","Data":"7dc3378ac936b4ff8d85d801f595353e896cb0d7d88d141b3d07f19be0ae7571"} Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.361639 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-volume-volume1-0" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.362811 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cindere6fe-account-delete-57xck" event={"ID":"9d764fb3-601e-4159-af7b-75c833e2792f","Type":"ContainerStarted","Data":"d647de38b479270292e82d6d4c55e47b8449dd4756f683dc64630857abdfead1"} Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.362930 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cindere6fe-account-delete-57xck" event={"ID":"9d764fb3-601e-4159-af7b-75c833e2792f","Type":"ContainerStarted","Data":"70a0cafda8e5be0054b3e384b2c8eba4b4f1d7829677d6f09e0bc00a614581cd"} Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.380395 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cinder-kuttl-tests/cindere6fe-account-delete-57xck" podStartSLOduration=1.380366421 podStartE2EDuration="1.380366421s" podCreationTimestamp="2026-03-11 13:43:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:43:33.377020493 +0000 UTC m=+1261.058922217" watchObservedRunningTime="2026-03-11 13:43:33.380366421 +0000 UTC m=+1261.062268135" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.381909 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-config-data" (OuterVolumeSpecName: "config-data") pod "722b7fb9-8dea-4dfb-9d69-d73216d237a6" (UID: "722b7fb9-8dea-4dfb-9d69-d73216d237a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.384838 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.384869 4923 reconciler_common.go:293] "Volume detached for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-var-locks-cinder\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.384881 4923 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-var-locks-brick\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.384894 4923 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.384904 4923 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.384915 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-config-data\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.384926 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkgx8\" (UniqueName: \"kubernetes.io/projected/722b7fb9-8dea-4dfb-9d69-d73216d237a6-kube-api-access-vkgx8\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.384938 4923 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-etc-nvme\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.384948 4923 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-sys\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.384961 4923 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-dev\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.384970 4923 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/722b7fb9-8dea-4dfb-9d69-d73216d237a6-run\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.384980 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/722b7fb9-8dea-4dfb-9d69-d73216d237a6-scripts\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.539110 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.540539 4923 scope.go:117] "RemoveContainer" containerID="2986755d10efa4502902211ac0a8fd41a81ad8db127f9a415998952f7619cfc1" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.546443 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.552954 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-volume-volume1-0"] Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.575049 4923 scope.go:117] "RemoveContainer" containerID="b6c5c5484239241499cb9d89d3783a167a592b3af17fa67e8c775ae590ce7c43" Mar 11 13:43:33 crc kubenswrapper[4923]: E0311 13:43:33.577106 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6c5c5484239241499cb9d89d3783a167a592b3af17fa67e8c775ae590ce7c43\": container with ID starting with b6c5c5484239241499cb9d89d3783a167a592b3af17fa67e8c775ae590ce7c43 not found: ID does not exist" containerID="b6c5c5484239241499cb9d89d3783a167a592b3af17fa67e8c775ae590ce7c43" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.577149 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6c5c5484239241499cb9d89d3783a167a592b3af17fa67e8c775ae590ce7c43"} err="failed to get container status \"b6c5c5484239241499cb9d89d3783a167a592b3af17fa67e8c775ae590ce7c43\": rpc error: code = NotFound desc = could not find container \"b6c5c5484239241499cb9d89d3783a167a592b3af17fa67e8c775ae590ce7c43\": container with ID starting with b6c5c5484239241499cb9d89d3783a167a592b3af17fa67e8c775ae590ce7c43 not found: ID does not exist" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.577177 4923 scope.go:117] "RemoveContainer" containerID="2986755d10efa4502902211ac0a8fd41a81ad8db127f9a415998952f7619cfc1" Mar 11 13:43:33 crc kubenswrapper[4923]: E0311 13:43:33.579906 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2986755d10efa4502902211ac0a8fd41a81ad8db127f9a415998952f7619cfc1\": container with ID starting with 2986755d10efa4502902211ac0a8fd41a81ad8db127f9a415998952f7619cfc1 not found: ID does not exist" containerID="2986755d10efa4502902211ac0a8fd41a81ad8db127f9a415998952f7619cfc1" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.579963 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2986755d10efa4502902211ac0a8fd41a81ad8db127f9a415998952f7619cfc1"} err="failed to get container status \"2986755d10efa4502902211ac0a8fd41a81ad8db127f9a415998952f7619cfc1\": rpc error: code = NotFound desc = could not find container \"2986755d10efa4502902211ac0a8fd41a81ad8db127f9a415998952f7619cfc1\": container with ID starting with 2986755d10efa4502902211ac0a8fd41a81ad8db127f9a415998952f7619cfc1 not found: ID does not exist" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.579982 4923 scope.go:117] "RemoveContainer" containerID="b6c5c5484239241499cb9d89d3783a167a592b3af17fa67e8c775ae590ce7c43" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.580283 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6c5c5484239241499cb9d89d3783a167a592b3af17fa67e8c775ae590ce7c43"} err="failed to get container status \"b6c5c5484239241499cb9d89d3783a167a592b3af17fa67e8c775ae590ce7c43\": rpc error: code = NotFound desc = could not find container \"b6c5c5484239241499cb9d89d3783a167a592b3af17fa67e8c775ae590ce7c43\": container with ID starting with b6c5c5484239241499cb9d89d3783a167a592b3af17fa67e8c775ae590ce7c43 not found: ID does not exist" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.580304 4923 scope.go:117] "RemoveContainer" containerID="2986755d10efa4502902211ac0a8fd41a81ad8db127f9a415998952f7619cfc1" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.580692 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2986755d10efa4502902211ac0a8fd41a81ad8db127f9a415998952f7619cfc1"} err="failed to get container status \"2986755d10efa4502902211ac0a8fd41a81ad8db127f9a415998952f7619cfc1\": rpc error: code = NotFound desc = could not find container \"2986755d10efa4502902211ac0a8fd41a81ad8db127f9a415998952f7619cfc1\": container with ID starting with 2986755d10efa4502902211ac0a8fd41a81ad8db127f9a415998952f7619cfc1 not found: ID does not exist" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.580720 4923 scope.go:117] "RemoveContainer" containerID="4f323cc06c832423a935bd57211a21d27ae7dfdd630767d0c158db5de4fb1394" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.594767 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zqvf\" (UniqueName: \"kubernetes.io/projected/12c7d96d-0e1a-4059-9e54-2a8438917bad-kube-api-access-9zqvf\") pod \"12c7d96d-0e1a-4059-9e54-2a8438917bad\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.594833 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12c7d96d-0e1a-4059-9e54-2a8438917bad-etc-machine-id\") pod \"12c7d96d-0e1a-4059-9e54-2a8438917bad\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.594872 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-scripts\") pod \"12c7d96d-0e1a-4059-9e54-2a8438917bad\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.594897 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-config-data-custom\") pod \"12c7d96d-0e1a-4059-9e54-2a8438917bad\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.594923 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-config-data\") pod \"12c7d96d-0e1a-4059-9e54-2a8438917bad\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.594971 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-combined-ca-bundle\") pod \"12c7d96d-0e1a-4059-9e54-2a8438917bad\" (UID: \"12c7d96d-0e1a-4059-9e54-2a8438917bad\") " Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.602732 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-scripts" (OuterVolumeSpecName: "scripts") pod "12c7d96d-0e1a-4059-9e54-2a8438917bad" (UID: "12c7d96d-0e1a-4059-9e54-2a8438917bad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.605730 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12c7d96d-0e1a-4059-9e54-2a8438917bad-kube-api-access-9zqvf" (OuterVolumeSpecName: "kube-api-access-9zqvf") pod "12c7d96d-0e1a-4059-9e54-2a8438917bad" (UID: "12c7d96d-0e1a-4059-9e54-2a8438917bad"). InnerVolumeSpecName "kube-api-access-9zqvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.605788 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/12c7d96d-0e1a-4059-9e54-2a8438917bad-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "12c7d96d-0e1a-4059-9e54-2a8438917bad" (UID: "12c7d96d-0e1a-4059-9e54-2a8438917bad"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.614505 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "12c7d96d-0e1a-4059-9e54-2a8438917bad" (UID: "12c7d96d-0e1a-4059-9e54-2a8438917bad"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.622273 4923 scope.go:117] "RemoveContainer" containerID="8bf10a9ec7b294d63184d7c43bf1a0508fd8b54697b5139a5716aae8a4ba506c" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.663012 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "12c7d96d-0e1a-4059-9e54-2a8438917bad" (UID: "12c7d96d-0e1a-4059-9e54-2a8438917bad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.684276 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.691149 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-backup-0"] Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.696044 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-scripts\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.696077 4923 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.696090 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.696103 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zqvf\" (UniqueName: \"kubernetes.io/projected/12c7d96d-0e1a-4059-9e54-2a8438917bad-kube-api-access-9zqvf\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.696116 4923 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/12c7d96d-0e1a-4059-9e54-2a8438917bad-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.697983 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-config-data" (OuterVolumeSpecName: "config-data") pod "12c7d96d-0e1a-4059-9e54-2a8438917bad" (UID: "12c7d96d-0e1a-4059-9e54-2a8438917bad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:43:33 crc kubenswrapper[4923]: I0311 13:43:33.797292 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/12c7d96d-0e1a-4059-9e54-2a8438917bad-config-data\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:34 crc kubenswrapper[4923]: I0311 13:43:34.373927 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-scheduler-0" event={"ID":"12c7d96d-0e1a-4059-9e54-2a8438917bad","Type":"ContainerDied","Data":"161c4c9a7f81dd4ee3a442d19a38488aba85b8beba4e404eb1f6d36c76b3c73a"} Mar 11 13:43:34 crc kubenswrapper[4923]: I0311 13:43:34.373979 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-scheduler-0" Mar 11 13:43:34 crc kubenswrapper[4923]: I0311 13:43:34.374334 4923 scope.go:117] "RemoveContainer" containerID="f6c9cd39e8280d1ff398def00133a920696fff51cced1383a3a022c0bd2d2284" Mar 11 13:43:34 crc kubenswrapper[4923]: I0311 13:43:34.376763 4923 generic.go:334] "Generic (PLEG): container finished" podID="9d764fb3-601e-4159-af7b-75c833e2792f" containerID="d647de38b479270292e82d6d4c55e47b8449dd4756f683dc64630857abdfead1" exitCode=0 Mar 11 13:43:34 crc kubenswrapper[4923]: I0311 13:43:34.376824 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cindere6fe-account-delete-57xck" event={"ID":"9d764fb3-601e-4159-af7b-75c833e2792f","Type":"ContainerDied","Data":"d647de38b479270292e82d6d4c55e47b8449dd4756f683dc64630857abdfead1"} Mar 11 13:43:34 crc kubenswrapper[4923]: I0311 13:43:34.396152 4923 scope.go:117] "RemoveContainer" containerID="58a54cd9a9104dcb279c12e8420fff8bf4b9f16614f8d75171a520ffa931f350" Mar 11 13:43:34 crc kubenswrapper[4923]: I0311 13:43:34.433760 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Mar 11 13:43:34 crc kubenswrapper[4923]: I0311 13:43:34.447690 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-scheduler-0"] Mar 11 13:43:35 crc kubenswrapper[4923]: I0311 13:43:35.038411 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12c7d96d-0e1a-4059-9e54-2a8438917bad" path="/var/lib/kubelet/pods/12c7d96d-0e1a-4059-9e54-2a8438917bad/volumes" Mar 11 13:43:35 crc kubenswrapper[4923]: I0311 13:43:35.039869 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="722b7fb9-8dea-4dfb-9d69-d73216d237a6" path="/var/lib/kubelet/pods/722b7fb9-8dea-4dfb-9d69-d73216d237a6/volumes" Mar 11 13:43:35 crc kubenswrapper[4923]: I0311 13:43:35.041086 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" path="/var/lib/kubelet/pods/ed2ffa6a-46e2-4257-a8c0-54d2808b37c4/volumes" Mar 11 13:43:35 crc kubenswrapper[4923]: I0311 13:43:35.784802 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cindere6fe-account-delete-57xck" Mar 11 13:43:35 crc kubenswrapper[4923]: I0311 13:43:35.877015 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:43:35 crc kubenswrapper[4923]: I0311 13:43:35.934013 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d764fb3-601e-4159-af7b-75c833e2792f-operator-scripts\") pod \"9d764fb3-601e-4159-af7b-75c833e2792f\" (UID: \"9d764fb3-601e-4159-af7b-75c833e2792f\") " Mar 11 13:43:35 crc kubenswrapper[4923]: I0311 13:43:35.934072 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9p8r\" (UniqueName: \"kubernetes.io/projected/9d764fb3-601e-4159-af7b-75c833e2792f-kube-api-access-n9p8r\") pod \"9d764fb3-601e-4159-af7b-75c833e2792f\" (UID: \"9d764fb3-601e-4159-af7b-75c833e2792f\") " Mar 11 13:43:35 crc kubenswrapper[4923]: I0311 13:43:35.935079 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d764fb3-601e-4159-af7b-75c833e2792f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9d764fb3-601e-4159-af7b-75c833e2792f" (UID: "9d764fb3-601e-4159-af7b-75c833e2792f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:43:35 crc kubenswrapper[4923]: I0311 13:43:35.938838 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d764fb3-601e-4159-af7b-75c833e2792f-kube-api-access-n9p8r" (OuterVolumeSpecName: "kube-api-access-n9p8r") pod "9d764fb3-601e-4159-af7b-75c833e2792f" (UID: "9d764fb3-601e-4159-af7b-75c833e2792f"). InnerVolumeSpecName "kube-api-access-n9p8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.035178 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2cf200ea-5fc3-4521-9d04-8662bc6783e7-etc-machine-id\") pod \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.035271 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-public-tls-certs\") pod \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.035300 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2cf200ea-5fc3-4521-9d04-8662bc6783e7-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2cf200ea-5fc3-4521-9d04-8662bc6783e7" (UID: "2cf200ea-5fc3-4521-9d04-8662bc6783e7"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.035336 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-scripts\") pod \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.035540 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-combined-ca-bundle\") pod \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.035649 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-internal-tls-certs\") pod \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.035721 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-config-data-custom\") pod \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.035810 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-config-data\") pod \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.035910 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2cf200ea-5fc3-4521-9d04-8662bc6783e7-logs\") pod \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.036937 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqgrj\" (UniqueName: \"kubernetes.io/projected/2cf200ea-5fc3-4521-9d04-8662bc6783e7-kube-api-access-lqgrj\") pod \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\" (UID: \"2cf200ea-5fc3-4521-9d04-8662bc6783e7\") " Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.036979 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2cf200ea-5fc3-4521-9d04-8662bc6783e7-logs" (OuterVolumeSpecName: "logs") pod "2cf200ea-5fc3-4521-9d04-8662bc6783e7" (UID: "2cf200ea-5fc3-4521-9d04-8662bc6783e7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.037455 4923 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d764fb3-601e-4159-af7b-75c833e2792f-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.037489 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9p8r\" (UniqueName: \"kubernetes.io/projected/9d764fb3-601e-4159-af7b-75c833e2792f-kube-api-access-n9p8r\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.037514 4923 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2cf200ea-5fc3-4521-9d04-8662bc6783e7-logs\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.037533 4923 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2cf200ea-5fc3-4521-9d04-8662bc6783e7-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.041693 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2cf200ea-5fc3-4521-9d04-8662bc6783e7" (UID: "2cf200ea-5fc3-4521-9d04-8662bc6783e7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.041799 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-scripts" (OuterVolumeSpecName: "scripts") pod "2cf200ea-5fc3-4521-9d04-8662bc6783e7" (UID: "2cf200ea-5fc3-4521-9d04-8662bc6783e7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.042510 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cf200ea-5fc3-4521-9d04-8662bc6783e7-kube-api-access-lqgrj" (OuterVolumeSpecName: "kube-api-access-lqgrj") pod "2cf200ea-5fc3-4521-9d04-8662bc6783e7" (UID: "2cf200ea-5fc3-4521-9d04-8662bc6783e7"). InnerVolumeSpecName "kube-api-access-lqgrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.066247 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2cf200ea-5fc3-4521-9d04-8662bc6783e7" (UID: "2cf200ea-5fc3-4521-9d04-8662bc6783e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.084182 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2cf200ea-5fc3-4521-9d04-8662bc6783e7" (UID: "2cf200ea-5fc3-4521-9d04-8662bc6783e7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.093285 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2cf200ea-5fc3-4521-9d04-8662bc6783e7" (UID: "2cf200ea-5fc3-4521-9d04-8662bc6783e7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.099443 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-config-data" (OuterVolumeSpecName: "config-data") pod "2cf200ea-5fc3-4521-9d04-8662bc6783e7" (UID: "2cf200ea-5fc3-4521-9d04-8662bc6783e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.139302 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.139374 4923 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.139395 4923 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.139413 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-config-data\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.139432 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqgrj\" (UniqueName: \"kubernetes.io/projected/2cf200ea-5fc3-4521-9d04-8662bc6783e7-kube-api-access-lqgrj\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.139452 4923 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.139469 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2cf200ea-5fc3-4521-9d04-8662bc6783e7-scripts\") on node \"crc\" DevicePath \"\"" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.400768 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cindere6fe-account-delete-57xck" event={"ID":"9d764fb3-601e-4159-af7b-75c833e2792f","Type":"ContainerDied","Data":"70a0cafda8e5be0054b3e384b2c8eba4b4f1d7829677d6f09e0bc00a614581cd"} Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.400837 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70a0cafda8e5be0054b3e384b2c8eba4b4f1d7829677d6f09e0bc00a614581cd" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.401427 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cindere6fe-account-delete-57xck" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.404562 4923 generic.go:334] "Generic (PLEG): container finished" podID="2cf200ea-5fc3-4521-9d04-8662bc6783e7" containerID="93bfb560b4a2dffc02b367035772405d304824390a522eaf35a2dafee7c5ab69" exitCode=0 Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.404618 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"2cf200ea-5fc3-4521-9d04-8662bc6783e7","Type":"ContainerDied","Data":"93bfb560b4a2dffc02b367035772405d304824390a522eaf35a2dafee7c5ab69"} Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.404650 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="cinder-kuttl-tests/cinder-api-0" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.404668 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cinder-kuttl-tests/cinder-api-0" event={"ID":"2cf200ea-5fc3-4521-9d04-8662bc6783e7","Type":"ContainerDied","Data":"e646f158c5ba329d6986b91d6f562258f9bbd6e977c65ec4075c9fa22dd8a13b"} Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.404708 4923 scope.go:117] "RemoveContainer" containerID="93bfb560b4a2dffc02b367035772405d304824390a522eaf35a2dafee7c5ab69" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.440438 4923 scope.go:117] "RemoveContainer" containerID="704e6c4e48c5154d167ecabca60342e620e915fdadad03f457a6ba3f0d8efffd" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.473120 4923 scope.go:117] "RemoveContainer" containerID="93bfb560b4a2dffc02b367035772405d304824390a522eaf35a2dafee7c5ab69" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.473297 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Mar 11 13:43:36 crc kubenswrapper[4923]: E0311 13:43:36.474242 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93bfb560b4a2dffc02b367035772405d304824390a522eaf35a2dafee7c5ab69\": container with ID starting with 93bfb560b4a2dffc02b367035772405d304824390a522eaf35a2dafee7c5ab69 not found: ID does not exist" containerID="93bfb560b4a2dffc02b367035772405d304824390a522eaf35a2dafee7c5ab69" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.474372 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93bfb560b4a2dffc02b367035772405d304824390a522eaf35a2dafee7c5ab69"} err="failed to get container status \"93bfb560b4a2dffc02b367035772405d304824390a522eaf35a2dafee7c5ab69\": rpc error: code = NotFound desc = could not find container \"93bfb560b4a2dffc02b367035772405d304824390a522eaf35a2dafee7c5ab69\": container with ID starting with 93bfb560b4a2dffc02b367035772405d304824390a522eaf35a2dafee7c5ab69 not found: ID does not exist" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.474420 4923 scope.go:117] "RemoveContainer" containerID="704e6c4e48c5154d167ecabca60342e620e915fdadad03f457a6ba3f0d8efffd" Mar 11 13:43:36 crc kubenswrapper[4923]: E0311 13:43:36.475000 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"704e6c4e48c5154d167ecabca60342e620e915fdadad03f457a6ba3f0d8efffd\": container with ID starting with 704e6c4e48c5154d167ecabca60342e620e915fdadad03f457a6ba3f0d8efffd not found: ID does not exist" containerID="704e6c4e48c5154d167ecabca60342e620e915fdadad03f457a6ba3f0d8efffd" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.475096 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"704e6c4e48c5154d167ecabca60342e620e915fdadad03f457a6ba3f0d8efffd"} err="failed to get container status \"704e6c4e48c5154d167ecabca60342e620e915fdadad03f457a6ba3f0d8efffd\": rpc error: code = NotFound desc = could not find container \"704e6c4e48c5154d167ecabca60342e620e915fdadad03f457a6ba3f0d8efffd\": container with ID starting with 704e6c4e48c5154d167ecabca60342e620e915fdadad03f457a6ba3f0d8efffd not found: ID does not exist" Mar 11 13:43:36 crc kubenswrapper[4923]: I0311 13:43:36.489043 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-api-0"] Mar 11 13:43:37 crc kubenswrapper[4923]: I0311 13:43:37.035313 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cf200ea-5fc3-4521-9d04-8662bc6783e7" path="/var/lib/kubelet/pods/2cf200ea-5fc3-4521-9d04-8662bc6783e7/volumes" Mar 11 13:43:37 crc kubenswrapper[4923]: I0311 13:43:37.204423 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cindere6fe-account-delete-57xck"] Mar 11 13:43:37 crc kubenswrapper[4923]: I0311 13:43:37.212707 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-24dq4"] Mar 11 13:43:37 crc kubenswrapper[4923]: I0311 13:43:37.222138 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cindere6fe-account-delete-57xck"] Mar 11 13:43:37 crc kubenswrapper[4923]: I0311 13:43:37.229177 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/cinder-e6fe-account-create-update-w2s4t"] Mar 11 13:43:37 crc kubenswrapper[4923]: I0311 13:43:37.234668 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-db-create-24dq4"] Mar 11 13:43:37 crc kubenswrapper[4923]: I0311 13:43:37.240220 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/cinder-e6fe-account-create-update-w2s4t"] Mar 11 13:43:39 crc kubenswrapper[4923]: I0311 13:43:39.041884 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6" path="/var/lib/kubelet/pods/1b2f3f3b-b87d-4a10-adfb-38cbdb70b2c6/volumes" Mar 11 13:43:39 crc kubenswrapper[4923]: I0311 13:43:39.042957 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d764fb3-601e-4159-af7b-75c833e2792f" path="/var/lib/kubelet/pods/9d764fb3-601e-4159-af7b-75c833e2792f/volumes" Mar 11 13:43:39 crc kubenswrapper[4923]: I0311 13:43:39.043608 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da36884b-e7cc-4ba1-9966-79b1a14059da" path="/var/lib/kubelet/pods/da36884b-e7cc-4ba1-9966-79b1a14059da/volumes" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.162077 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29553944-hktnj"] Mar 11 13:44:00 crc kubenswrapper[4923]: E0311 13:44:00.163437 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="cinder-volume" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.163466 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="cinder-volume" Mar 11 13:44:00 crc kubenswrapper[4923]: E0311 13:44:00.163489 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c7d96d-0e1a-4059-9e54-2a8438917bad" containerName="cinder-scheduler" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.163505 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c7d96d-0e1a-4059-9e54-2a8438917bad" containerName="cinder-scheduler" Mar 11 13:44:00 crc kubenswrapper[4923]: E0311 13:44:00.163524 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="cinder-volume" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.163540 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="cinder-volume" Mar 11 13:44:00 crc kubenswrapper[4923]: E0311 13:44:00.163564 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="cinder-volume" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.163580 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="cinder-volume" Mar 11 13:44:00 crc kubenswrapper[4923]: E0311 13:44:00.163600 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="probe" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.163615 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="probe" Mar 11 13:44:00 crc kubenswrapper[4923]: E0311 13:44:00.163633 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cf200ea-5fc3-4521-9d04-8662bc6783e7" containerName="cinder-api" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.163649 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cf200ea-5fc3-4521-9d04-8662bc6783e7" containerName="cinder-api" Mar 11 13:44:00 crc kubenswrapper[4923]: E0311 13:44:00.163668 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cf200ea-5fc3-4521-9d04-8662bc6783e7" containerName="cinder-api-log" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.163684 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cf200ea-5fc3-4521-9d04-8662bc6783e7" containerName="cinder-api-log" Mar 11 13:44:00 crc kubenswrapper[4923]: E0311 13:44:00.163717 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="722b7fb9-8dea-4dfb-9d69-d73216d237a6" containerName="cinder-backup" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.163901 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="722b7fb9-8dea-4dfb-9d69-d73216d237a6" containerName="cinder-backup" Mar 11 13:44:00 crc kubenswrapper[4923]: E0311 13:44:00.163920 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="722b7fb9-8dea-4dfb-9d69-d73216d237a6" containerName="probe" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.163938 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="722b7fb9-8dea-4dfb-9d69-d73216d237a6" containerName="probe" Mar 11 13:44:00 crc kubenswrapper[4923]: E0311 13:44:00.163965 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="probe" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.163980 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="probe" Mar 11 13:44:00 crc kubenswrapper[4923]: E0311 13:44:00.164004 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="cinder-volume" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.164019 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="cinder-volume" Mar 11 13:44:00 crc kubenswrapper[4923]: E0311 13:44:00.164038 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="probe" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.164055 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="probe" Mar 11 13:44:00 crc kubenswrapper[4923]: E0311 13:44:00.164076 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d764fb3-601e-4159-af7b-75c833e2792f" containerName="mariadb-account-delete" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.164091 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d764fb3-601e-4159-af7b-75c833e2792f" containerName="mariadb-account-delete" Mar 11 13:44:00 crc kubenswrapper[4923]: E0311 13:44:00.164110 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="probe" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.164126 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="probe" Mar 11 13:44:00 crc kubenswrapper[4923]: E0311 13:44:00.164149 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c7d96d-0e1a-4059-9e54-2a8438917bad" containerName="probe" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.164166 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c7d96d-0e1a-4059-9e54-2a8438917bad" containerName="probe" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.164468 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cf200ea-5fc3-4521-9d04-8662bc6783e7" containerName="cinder-api" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.164501 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="cinder-volume" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.164518 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d764fb3-601e-4159-af7b-75c833e2792f" containerName="mariadb-account-delete" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.164545 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="cinder-volume" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.164565 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="722b7fb9-8dea-4dfb-9d69-d73216d237a6" containerName="probe" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.164594 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="probe" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.164612 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="12c7d96d-0e1a-4059-9e54-2a8438917bad" containerName="cinder-scheduler" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.164635 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cf200ea-5fc3-4521-9d04-8662bc6783e7" containerName="cinder-api-log" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.164661 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="12c7d96d-0e1a-4059-9e54-2a8438917bad" containerName="probe" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.164693 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="probe" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.164709 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="probe" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.164729 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="probe" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.164751 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="cinder-volume" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.164770 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="722b7fb9-8dea-4dfb-9d69-d73216d237a6" containerName="cinder-backup" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.165717 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553944-hktnj" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.167887 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553944-hktnj"] Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.169130 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-nxck7" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.169507 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.170851 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.258598 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mhrs\" (UniqueName: \"kubernetes.io/projected/9e956abd-be22-488a-895d-171f5ebe495b-kube-api-access-5mhrs\") pod \"auto-csr-approver-29553944-hktnj\" (UID: \"9e956abd-be22-488a-895d-171f5ebe495b\") " pod="openshift-infra/auto-csr-approver-29553944-hktnj" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.360173 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mhrs\" (UniqueName: \"kubernetes.io/projected/9e956abd-be22-488a-895d-171f5ebe495b-kube-api-access-5mhrs\") pod \"auto-csr-approver-29553944-hktnj\" (UID: \"9e956abd-be22-488a-895d-171f5ebe495b\") " pod="openshift-infra/auto-csr-approver-29553944-hktnj" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.394654 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mhrs\" (UniqueName: \"kubernetes.io/projected/9e956abd-be22-488a-895d-171f5ebe495b-kube-api-access-5mhrs\") pod \"auto-csr-approver-29553944-hktnj\" (UID: \"9e956abd-be22-488a-895d-171f5ebe495b\") " pod="openshift-infra/auto-csr-approver-29553944-hktnj" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.500322 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553944-hktnj" Mar 11 13:44:00 crc kubenswrapper[4923]: I0311 13:44:00.723760 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553944-hktnj"] Mar 11 13:44:01 crc kubenswrapper[4923]: I0311 13:44:01.651072 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553944-hktnj" event={"ID":"9e956abd-be22-488a-895d-171f5ebe495b","Type":"ContainerStarted","Data":"b39c65974dbb6a4fd5b8d722a1961b122d2dbf2e949c51b14dd1583efb1b1ff4"} Mar 11 13:44:02 crc kubenswrapper[4923]: I0311 13:44:02.664556 4923 generic.go:334] "Generic (PLEG): container finished" podID="9e956abd-be22-488a-895d-171f5ebe495b" containerID="4f86578704f9bdee6ff6aa583785d75e6d018c9cb4f8cf5cd6f59c34460769ce" exitCode=0 Mar 11 13:44:02 crc kubenswrapper[4923]: I0311 13:44:02.664879 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553944-hktnj" event={"ID":"9e956abd-be22-488a-895d-171f5ebe495b","Type":"ContainerDied","Data":"4f86578704f9bdee6ff6aa583785d75e6d018c9cb4f8cf5cd6f59c34460769ce"} Mar 11 13:44:04 crc kubenswrapper[4923]: I0311 13:44:04.826014 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553944-hktnj" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.024458 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5mhrs\" (UniqueName: \"kubernetes.io/projected/9e956abd-be22-488a-895d-171f5ebe495b-kube-api-access-5mhrs\") pod \"9e956abd-be22-488a-895d-171f5ebe495b\" (UID: \"9e956abd-be22-488a-895d-171f5ebe495b\") " Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.032289 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e956abd-be22-488a-895d-171f5ebe495b-kube-api-access-5mhrs" (OuterVolumeSpecName: "kube-api-access-5mhrs") pod "9e956abd-be22-488a-895d-171f5ebe495b" (UID: "9e956abd-be22-488a-895d-171f5ebe495b"). InnerVolumeSpecName "kube-api-access-5mhrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.126598 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5mhrs\" (UniqueName: \"kubernetes.io/projected/9e956abd-be22-488a-895d-171f5ebe495b-kube-api-access-5mhrs\") on node \"crc\" DevicePath \"\"" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.269711 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-h9dp2/must-gather-fn4dg"] Mar 11 13:44:05 crc kubenswrapper[4923]: E0311 13:44:05.269962 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="probe" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.269979 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="probe" Mar 11 13:44:05 crc kubenswrapper[4923]: E0311 13:44:05.270003 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="cinder-volume" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.270010 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="cinder-volume" Mar 11 13:44:05 crc kubenswrapper[4923]: E0311 13:44:05.270023 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e956abd-be22-488a-895d-171f5ebe495b" containerName="oc" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.270029 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e956abd-be22-488a-895d-171f5ebe495b" containerName="oc" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.270143 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e956abd-be22-488a-895d-171f5ebe495b" containerName="oc" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.270167 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="probe" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.270179 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="cinder-volume" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.270188 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed2ffa6a-46e2-4257-a8c0-54d2808b37c4" containerName="cinder-volume" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.270859 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h9dp2/must-gather-fn4dg" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.274778 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-h9dp2"/"default-dockercfg-m95l2" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.276977 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-h9dp2"/"kube-root-ca.crt" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.277700 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-h9dp2"/"openshift-service-ca.crt" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.279323 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-h9dp2/must-gather-fn4dg"] Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.430333 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/931f29a0-a23b-4191-9abd-19026cb35843-must-gather-output\") pod \"must-gather-fn4dg\" (UID: \"931f29a0-a23b-4191-9abd-19026cb35843\") " pod="openshift-must-gather-h9dp2/must-gather-fn4dg" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.430449 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmgkj\" (UniqueName: \"kubernetes.io/projected/931f29a0-a23b-4191-9abd-19026cb35843-kube-api-access-xmgkj\") pod \"must-gather-fn4dg\" (UID: \"931f29a0-a23b-4191-9abd-19026cb35843\") " pod="openshift-must-gather-h9dp2/must-gather-fn4dg" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.532028 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/931f29a0-a23b-4191-9abd-19026cb35843-must-gather-output\") pod \"must-gather-fn4dg\" (UID: \"931f29a0-a23b-4191-9abd-19026cb35843\") " pod="openshift-must-gather-h9dp2/must-gather-fn4dg" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.532116 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmgkj\" (UniqueName: \"kubernetes.io/projected/931f29a0-a23b-4191-9abd-19026cb35843-kube-api-access-xmgkj\") pod \"must-gather-fn4dg\" (UID: \"931f29a0-a23b-4191-9abd-19026cb35843\") " pod="openshift-must-gather-h9dp2/must-gather-fn4dg" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.532515 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/931f29a0-a23b-4191-9abd-19026cb35843-must-gather-output\") pod \"must-gather-fn4dg\" (UID: \"931f29a0-a23b-4191-9abd-19026cb35843\") " pod="openshift-must-gather-h9dp2/must-gather-fn4dg" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.560674 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553944-hktnj" event={"ID":"9e956abd-be22-488a-895d-171f5ebe495b","Type":"ContainerDied","Data":"b39c65974dbb6a4fd5b8d722a1961b122d2dbf2e949c51b14dd1583efb1b1ff4"} Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.561079 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b39c65974dbb6a4fd5b8d722a1961b122d2dbf2e949c51b14dd1583efb1b1ff4" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.560712 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553944-hktnj" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.566149 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmgkj\" (UniqueName: \"kubernetes.io/projected/931f29a0-a23b-4191-9abd-19026cb35843-kube-api-access-xmgkj\") pod \"must-gather-fn4dg\" (UID: \"931f29a0-a23b-4191-9abd-19026cb35843\") " pod="openshift-must-gather-h9dp2/must-gather-fn4dg" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.585386 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h9dp2/must-gather-fn4dg" Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.903651 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29553938-tmvzr"] Mar 11 13:44:05 crc kubenswrapper[4923]: I0311 13:44:05.911485 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29553938-tmvzr"] Mar 11 13:44:06 crc kubenswrapper[4923]: I0311 13:44:06.048533 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-h9dp2/must-gather-fn4dg"] Mar 11 13:44:06 crc kubenswrapper[4923]: W0311 13:44:06.051581 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod931f29a0_a23b_4191_9abd_19026cb35843.slice/crio-f7d65279fbf719b2746648ed24d9731aaf33d8c0e78ffd44c507f06b365a6584 WatchSource:0}: Error finding container f7d65279fbf719b2746648ed24d9731aaf33d8c0e78ffd44c507f06b365a6584: Status 404 returned error can't find the container with id f7d65279fbf719b2746648ed24d9731aaf33d8c0e78ffd44c507f06b365a6584 Mar 11 13:44:06 crc kubenswrapper[4923]: I0311 13:44:06.575883 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h9dp2/must-gather-fn4dg" event={"ID":"931f29a0-a23b-4191-9abd-19026cb35843","Type":"ContainerStarted","Data":"f7d65279fbf719b2746648ed24d9731aaf33d8c0e78ffd44c507f06b365a6584"} Mar 11 13:44:07 crc kubenswrapper[4923]: I0311 13:44:07.032154 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91414f8a-2c8d-4aa0-8d82-3dd35377ef4c" path="/var/lib/kubelet/pods/91414f8a-2c8d-4aa0-8d82-3dd35377ef4c/volumes" Mar 11 13:44:11 crc kubenswrapper[4923]: I0311 13:44:11.932012 4923 patch_prober.go:28] interesting pod/machine-config-daemon-fsfnw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 11 13:44:11 crc kubenswrapper[4923]: I0311 13:44:11.932581 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 11 13:44:12 crc kubenswrapper[4923]: I0311 13:44:12.621253 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h9dp2/must-gather-fn4dg" event={"ID":"931f29a0-a23b-4191-9abd-19026cb35843","Type":"ContainerStarted","Data":"ce723073b20c7bb87e20d9abfbe598299afabe97f123e455b9c390f55398155a"} Mar 11 13:44:12 crc kubenswrapper[4923]: I0311 13:44:12.621613 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h9dp2/must-gather-fn4dg" event={"ID":"931f29a0-a23b-4191-9abd-19026cb35843","Type":"ContainerStarted","Data":"8ed5b7eb28064e58bae9d9bea8486d9ae035b6b14290d20e70c1106ef634bc8f"} Mar 11 13:44:34 crc kubenswrapper[4923]: I0311 13:44:34.234299 4923 scope.go:117] "RemoveContainer" containerID="f53ae113b8c07d010ea14334e7eb25e74c34712b50222a8106bbce8973c62c8a" Mar 11 13:44:41 crc kubenswrapper[4923]: I0311 13:44:41.931987 4923 patch_prober.go:28] interesting pod/machine-config-daemon-fsfnw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 11 13:44:41 crc kubenswrapper[4923]: I0311 13:44:41.932519 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 11 13:44:44 crc kubenswrapper[4923]: I0311 13:44:44.860440 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-h9dp2/must-gather-fn4dg" podStartSLOduration=33.892150215 podStartE2EDuration="39.86041426s" podCreationTimestamp="2026-03-11 13:44:05 +0000 UTC" firstStartedPulling="2026-03-11 13:44:06.05627211 +0000 UTC m=+1293.738173824" lastFinishedPulling="2026-03-11 13:44:12.024536155 +0000 UTC m=+1299.706437869" observedRunningTime="2026-03-11 13:44:12.649539369 +0000 UTC m=+1300.331441083" watchObservedRunningTime="2026-03-11 13:44:44.86041426 +0000 UTC m=+1332.542315984" Mar 11 13:44:44 crc kubenswrapper[4923]: I0311 13:44:44.864859 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9vqtj"] Mar 11 13:44:44 crc kubenswrapper[4923]: I0311 13:44:44.866765 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9vqtj" Mar 11 13:44:44 crc kubenswrapper[4923]: I0311 13:44:44.879670 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9vqtj"] Mar 11 13:44:44 crc kubenswrapper[4923]: I0311 13:44:44.927065 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d39512-98cd-454c-8100-4d58a594797b-utilities\") pod \"certified-operators-9vqtj\" (UID: \"b3d39512-98cd-454c-8100-4d58a594797b\") " pod="openshift-marketplace/certified-operators-9vqtj" Mar 11 13:44:44 crc kubenswrapper[4923]: I0311 13:44:44.927554 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xd47k\" (UniqueName: \"kubernetes.io/projected/b3d39512-98cd-454c-8100-4d58a594797b-kube-api-access-xd47k\") pod \"certified-operators-9vqtj\" (UID: \"b3d39512-98cd-454c-8100-4d58a594797b\") " pod="openshift-marketplace/certified-operators-9vqtj" Mar 11 13:44:44 crc kubenswrapper[4923]: I0311 13:44:44.927642 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d39512-98cd-454c-8100-4d58a594797b-catalog-content\") pod \"certified-operators-9vqtj\" (UID: \"b3d39512-98cd-454c-8100-4d58a594797b\") " pod="openshift-marketplace/certified-operators-9vqtj" Mar 11 13:44:45 crc kubenswrapper[4923]: I0311 13:44:45.028764 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xd47k\" (UniqueName: \"kubernetes.io/projected/b3d39512-98cd-454c-8100-4d58a594797b-kube-api-access-xd47k\") pod \"certified-operators-9vqtj\" (UID: \"b3d39512-98cd-454c-8100-4d58a594797b\") " pod="openshift-marketplace/certified-operators-9vqtj" Mar 11 13:44:45 crc kubenswrapper[4923]: I0311 13:44:45.028838 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d39512-98cd-454c-8100-4d58a594797b-catalog-content\") pod \"certified-operators-9vqtj\" (UID: \"b3d39512-98cd-454c-8100-4d58a594797b\") " pod="openshift-marketplace/certified-operators-9vqtj" Mar 11 13:44:45 crc kubenswrapper[4923]: I0311 13:44:45.028930 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d39512-98cd-454c-8100-4d58a594797b-utilities\") pod \"certified-operators-9vqtj\" (UID: \"b3d39512-98cd-454c-8100-4d58a594797b\") " pod="openshift-marketplace/certified-operators-9vqtj" Mar 11 13:44:45 crc kubenswrapper[4923]: I0311 13:44:45.029369 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d39512-98cd-454c-8100-4d58a594797b-utilities\") pod \"certified-operators-9vqtj\" (UID: \"b3d39512-98cd-454c-8100-4d58a594797b\") " pod="openshift-marketplace/certified-operators-9vqtj" Mar 11 13:44:45 crc kubenswrapper[4923]: I0311 13:44:45.029584 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d39512-98cd-454c-8100-4d58a594797b-catalog-content\") pod \"certified-operators-9vqtj\" (UID: \"b3d39512-98cd-454c-8100-4d58a594797b\") " pod="openshift-marketplace/certified-operators-9vqtj" Mar 11 13:44:45 crc kubenswrapper[4923]: I0311 13:44:45.054880 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xd47k\" (UniqueName: \"kubernetes.io/projected/b3d39512-98cd-454c-8100-4d58a594797b-kube-api-access-xd47k\") pod \"certified-operators-9vqtj\" (UID: \"b3d39512-98cd-454c-8100-4d58a594797b\") " pod="openshift-marketplace/certified-operators-9vqtj" Mar 11 13:44:45 crc kubenswrapper[4923]: I0311 13:44:45.184516 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9vqtj" Mar 11 13:44:45 crc kubenswrapper[4923]: I0311 13:44:45.637731 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9vqtj"] Mar 11 13:44:45 crc kubenswrapper[4923]: I0311 13:44:45.865763 4923 generic.go:334] "Generic (PLEG): container finished" podID="b3d39512-98cd-454c-8100-4d58a594797b" containerID="6ab41d0d442faa62ff70a87de2dc3417f01d6d13c0b7b1e3ea49b4407061ff36" exitCode=0 Mar 11 13:44:45 crc kubenswrapper[4923]: I0311 13:44:45.865804 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vqtj" event={"ID":"b3d39512-98cd-454c-8100-4d58a594797b","Type":"ContainerDied","Data":"6ab41d0d442faa62ff70a87de2dc3417f01d6d13c0b7b1e3ea49b4407061ff36"} Mar 11 13:44:45 crc kubenswrapper[4923]: I0311 13:44:45.865830 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vqtj" event={"ID":"b3d39512-98cd-454c-8100-4d58a594797b","Type":"ContainerStarted","Data":"e243b858aa912a7d77015009663d61725db6855c6fc527fa12c37ef14b93888a"} Mar 11 13:44:47 crc kubenswrapper[4923]: I0311 13:44:47.882121 4923 generic.go:334] "Generic (PLEG): container finished" podID="b3d39512-98cd-454c-8100-4d58a594797b" containerID="52369a91e56a0741363a176fa1fd5114f82a9a44636b7498c193a0fb10efe225" exitCode=0 Mar 11 13:44:47 crc kubenswrapper[4923]: I0311 13:44:47.882235 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vqtj" event={"ID":"b3d39512-98cd-454c-8100-4d58a594797b","Type":"ContainerDied","Data":"52369a91e56a0741363a176fa1fd5114f82a9a44636b7498c193a0fb10efe225"} Mar 11 13:44:48 crc kubenswrapper[4923]: I0311 13:44:48.890434 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vqtj" event={"ID":"b3d39512-98cd-454c-8100-4d58a594797b","Type":"ContainerStarted","Data":"423a991a4319b8501e3cb40adafe5dd62242132de34d1fab19ac2b3b1db77a86"} Mar 11 13:44:53 crc kubenswrapper[4923]: I0311 13:44:53.852771 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94_df8084e0-b75d-4c85-93ea-61cf2e3281b6/util/0.log" Mar 11 13:44:53 crc kubenswrapper[4923]: I0311 13:44:53.989934 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94_df8084e0-b75d-4c85-93ea-61cf2e3281b6/pull/0.log" Mar 11 13:44:54 crc kubenswrapper[4923]: I0311 13:44:54.006688 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94_df8084e0-b75d-4c85-93ea-61cf2e3281b6/util/0.log" Mar 11 13:44:54 crc kubenswrapper[4923]: I0311 13:44:54.027136 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94_df8084e0-b75d-4c85-93ea-61cf2e3281b6/pull/0.log" Mar 11 13:44:54 crc kubenswrapper[4923]: I0311 13:44:54.192778 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94_df8084e0-b75d-4c85-93ea-61cf2e3281b6/util/0.log" Mar 11 13:44:54 crc kubenswrapper[4923]: I0311 13:44:54.199216 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94_df8084e0-b75d-4c85-93ea-61cf2e3281b6/pull/0.log" Mar 11 13:44:54 crc kubenswrapper[4923]: I0311 13:44:54.209810 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590kfl94_df8084e0-b75d-4c85-93ea-61cf2e3281b6/extract/0.log" Mar 11 13:44:54 crc kubenswrapper[4923]: I0311 13:44:54.372185 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr_fc311cce-a1ef-45e8-be23-2b5c55bb4243/util/0.log" Mar 11 13:44:54 crc kubenswrapper[4923]: I0311 13:44:54.538621 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr_fc311cce-a1ef-45e8-be23-2b5c55bb4243/util/0.log" Mar 11 13:44:54 crc kubenswrapper[4923]: I0311 13:44:54.587281 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr_fc311cce-a1ef-45e8-be23-2b5c55bb4243/pull/0.log" Mar 11 13:44:54 crc kubenswrapper[4923]: I0311 13:44:54.602661 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr_fc311cce-a1ef-45e8-be23-2b5c55bb4243/pull/0.log" Mar 11 13:44:54 crc kubenswrapper[4923]: I0311 13:44:54.763665 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr_fc311cce-a1ef-45e8-be23-2b5c55bb4243/util/0.log" Mar 11 13:44:54 crc kubenswrapper[4923]: I0311 13:44:54.783376 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr_fc311cce-a1ef-45e8-be23-2b5c55bb4243/pull/0.log" Mar 11 13:44:54 crc kubenswrapper[4923]: I0311 13:44:54.800634 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a8831b5415c61d2070187f89bbd36b8acd1ec1f7bfd7a0222b24d4bc40k5lqr_fc311cce-a1ef-45e8-be23-2b5c55bb4243/extract/0.log" Mar 11 13:44:54 crc kubenswrapper[4923]: I0311 13:44:54.918214 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2_2726a877-fccf-4cd2-95a7-084920be7c21/util/0.log" Mar 11 13:44:55 crc kubenswrapper[4923]: I0311 13:44:55.105372 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2_2726a877-fccf-4cd2-95a7-084920be7c21/pull/0.log" Mar 11 13:44:55 crc kubenswrapper[4923]: I0311 13:44:55.109743 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2_2726a877-fccf-4cd2-95a7-084920be7c21/util/0.log" Mar 11 13:44:55 crc kubenswrapper[4923]: I0311 13:44:55.123825 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2_2726a877-fccf-4cd2-95a7-084920be7c21/pull/0.log" Mar 11 13:44:55 crc kubenswrapper[4923]: I0311 13:44:55.185710 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9vqtj" Mar 11 13:44:55 crc kubenswrapper[4923]: I0311 13:44:55.185756 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9vqtj" Mar 11 13:44:55 crc kubenswrapper[4923]: I0311 13:44:55.233322 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9vqtj" Mar 11 13:44:55 crc kubenswrapper[4923]: I0311 13:44:55.251619 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9vqtj" podStartSLOduration=8.642569798 podStartE2EDuration="11.25160111s" podCreationTimestamp="2026-03-11 13:44:44 +0000 UTC" firstStartedPulling="2026-03-11 13:44:45.86704853 +0000 UTC m=+1333.548950244" lastFinishedPulling="2026-03-11 13:44:48.476079842 +0000 UTC m=+1336.157981556" observedRunningTime="2026-03-11 13:44:48.914012868 +0000 UTC m=+1336.595914582" watchObservedRunningTime="2026-03-11 13:44:55.25160111 +0000 UTC m=+1342.933502834" Mar 11 13:44:55 crc kubenswrapper[4923]: I0311 13:44:55.303491 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2_2726a877-fccf-4cd2-95a7-084920be7c21/util/0.log" Mar 11 13:44:55 crc kubenswrapper[4923]: I0311 13:44:55.304291 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2_2726a877-fccf-4cd2-95a7-084920be7c21/pull/0.log" Mar 11 13:44:55 crc kubenswrapper[4923]: I0311 13:44:55.304432 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_b8fc154d892d0d2aa94a34c8600d1a0cab4cdca8abc09f645bfd1f1da65knh2_2726a877-fccf-4cd2-95a7-084920be7c21/extract/0.log" Mar 11 13:44:55 crc kubenswrapper[4923]: I0311 13:44:55.467945 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-index-66xkx_4c504c2c-5548-406f-b2f9-55f53c126a0f/registry-server/0.log" Mar 11 13:44:55 crc kubenswrapper[4923]: I0311 13:44:55.690516 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l_c039997d-8c74-456b-8d4e-250491cff9fc/util/0.log" Mar 11 13:44:55 crc kubenswrapper[4923]: I0311 13:44:55.903957 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-8698b9b776-6c84l_b2d6a790-42f2-4c33-a7d5-24e033961c36/manager/0.log" Mar 11 13:44:55 crc kubenswrapper[4923]: I0311 13:44:55.939431 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l_c039997d-8c74-456b-8d4e-250491cff9fc/util/0.log" Mar 11 13:44:55 crc kubenswrapper[4923]: I0311 13:44:55.961578 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l_c039997d-8c74-456b-8d4e-250491cff9fc/pull/0.log" Mar 11 13:44:56 crc kubenswrapper[4923]: I0311 13:44:56.010884 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9vqtj" Mar 11 13:44:56 crc kubenswrapper[4923]: I0311 13:44:56.061460 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9vqtj"] Mar 11 13:44:56 crc kubenswrapper[4923]: I0311 13:44:56.089992 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l_c039997d-8c74-456b-8d4e-250491cff9fc/pull/0.log" Mar 11 13:44:56 crc kubenswrapper[4923]: I0311 13:44:56.198302 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l_c039997d-8c74-456b-8d4e-250491cff9fc/pull/0.log" Mar 11 13:44:56 crc kubenswrapper[4923]: I0311 13:44:56.215186 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l_c039997d-8c74-456b-8d4e-250491cff9fc/extract/0.log" Mar 11 13:44:56 crc kubenswrapper[4923]: I0311 13:44:56.221526 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d402d75b3e5d6ff4ae549f240fa3a42df571ed09a4202c90d5fd954546dcj2l_c039997d-8c74-456b-8d4e-250491cff9fc/util/0.log" Mar 11 13:44:56 crc kubenswrapper[4923]: I0311 13:44:56.387727 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf_5ec59e42-a1e4-4190-95bc-8370d92be5ff/util/0.log" Mar 11 13:44:56 crc kubenswrapper[4923]: I0311 13:44:56.565715 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf_5ec59e42-a1e4-4190-95bc-8370d92be5ff/pull/0.log" Mar 11 13:44:56 crc kubenswrapper[4923]: I0311 13:44:56.580093 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf_5ec59e42-a1e4-4190-95bc-8370d92be5ff/util/0.log" Mar 11 13:44:56 crc kubenswrapper[4923]: I0311 13:44:56.620650 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf_5ec59e42-a1e4-4190-95bc-8370d92be5ff/pull/0.log" Mar 11 13:44:56 crc kubenswrapper[4923]: I0311 13:44:56.796314 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf_5ec59e42-a1e4-4190-95bc-8370d92be5ff/util/0.log" Mar 11 13:44:56 crc kubenswrapper[4923]: I0311 13:44:56.832902 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf_5ec59e42-a1e4-4190-95bc-8370d92be5ff/extract/0.log" Mar 11 13:44:56 crc kubenswrapper[4923]: I0311 13:44:56.837285 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ea64798c11a47584aa8ef9f79c6d74c5f8853c9434c0c526d0c053f225b5wkf_5ec59e42-a1e4-4190-95bc-8370d92be5ff/pull/0.log" Mar 11 13:44:56 crc kubenswrapper[4923]: I0311 13:44:56.990048 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-index-p8zpv_656e3de8-2700-4579-ad0f-961699ace32e/registry-server/0.log" Mar 11 13:44:57 crc kubenswrapper[4923]: I0311 13:44:57.000536 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5b8cc7847-7bm9z_88950f18-28e6-4e1f-92a3-c00047b542da/manager/0.log" Mar 11 13:44:57 crc kubenswrapper[4923]: I0311 13:44:57.202677 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-index-5jw2x_ff2988d3-36a1-4085-935a-6d9ab916faac/registry-server/0.log" Mar 11 13:44:57 crc kubenswrapper[4923]: I0311 13:44:57.229514 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7f7964dd58-bjc24_a492ddb6-a936-4f7d-ba13-d4ddb83343a0/manager/0.log" Mar 11 13:44:57 crc kubenswrapper[4923]: I0311 13:44:57.262327 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-6d9597b6cb-f8lmr_6755f934-96d2-4447-b003-aea03c8df883/manager/0.log" Mar 11 13:44:57 crc kubenswrapper[4923]: I0311 13:44:57.351000 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-index-btj29_8f258a09-0cb8-43c1-9f1b-72b75147353c/registry-server/0.log" Mar 11 13:44:57 crc kubenswrapper[4923]: I0311 13:44:57.414763 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-779fc9694b-5xpgn_eb048f52-cfe4-4827-9670-6406a758f39f/operator/0.log" Mar 11 13:44:57 crc kubenswrapper[4923]: I0311 13:44:57.530091 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-index-sd5l9_9e845d8a-4bf4-457c-a8c3-871ed232846d/registry-server/0.log" Mar 11 13:44:57 crc kubenswrapper[4923]: I0311 13:44:57.960537 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9vqtj" podUID="b3d39512-98cd-454c-8100-4d58a594797b" containerName="registry-server" containerID="cri-o://423a991a4319b8501e3cb40adafe5dd62242132de34d1fab19ac2b3b1db77a86" gracePeriod=2 Mar 11 13:44:58 crc kubenswrapper[4923]: I0311 13:44:58.970949 4923 generic.go:334] "Generic (PLEG): container finished" podID="b3d39512-98cd-454c-8100-4d58a594797b" containerID="423a991a4319b8501e3cb40adafe5dd62242132de34d1fab19ac2b3b1db77a86" exitCode=0 Mar 11 13:44:58 crc kubenswrapper[4923]: I0311 13:44:58.971029 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vqtj" event={"ID":"b3d39512-98cd-454c-8100-4d58a594797b","Type":"ContainerDied","Data":"423a991a4319b8501e3cb40adafe5dd62242132de34d1fab19ac2b3b1db77a86"} Mar 11 13:44:59 crc kubenswrapper[4923]: I0311 13:44:59.457061 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9vqtj" Mar 11 13:44:59 crc kubenswrapper[4923]: I0311 13:44:59.543681 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d39512-98cd-454c-8100-4d58a594797b-utilities\") pod \"b3d39512-98cd-454c-8100-4d58a594797b\" (UID: \"b3d39512-98cd-454c-8100-4d58a594797b\") " Mar 11 13:44:59 crc kubenswrapper[4923]: I0311 13:44:59.543723 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d39512-98cd-454c-8100-4d58a594797b-catalog-content\") pod \"b3d39512-98cd-454c-8100-4d58a594797b\" (UID: \"b3d39512-98cd-454c-8100-4d58a594797b\") " Mar 11 13:44:59 crc kubenswrapper[4923]: I0311 13:44:59.543819 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xd47k\" (UniqueName: \"kubernetes.io/projected/b3d39512-98cd-454c-8100-4d58a594797b-kube-api-access-xd47k\") pod \"b3d39512-98cd-454c-8100-4d58a594797b\" (UID: \"b3d39512-98cd-454c-8100-4d58a594797b\") " Mar 11 13:44:59 crc kubenswrapper[4923]: I0311 13:44:59.545505 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3d39512-98cd-454c-8100-4d58a594797b-utilities" (OuterVolumeSpecName: "utilities") pod "b3d39512-98cd-454c-8100-4d58a594797b" (UID: "b3d39512-98cd-454c-8100-4d58a594797b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:44:59 crc kubenswrapper[4923]: I0311 13:44:59.556621 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3d39512-98cd-454c-8100-4d58a594797b-kube-api-access-xd47k" (OuterVolumeSpecName: "kube-api-access-xd47k") pod "b3d39512-98cd-454c-8100-4d58a594797b" (UID: "b3d39512-98cd-454c-8100-4d58a594797b"). InnerVolumeSpecName "kube-api-access-xd47k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:44:59 crc kubenswrapper[4923]: I0311 13:44:59.601170 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3d39512-98cd-454c-8100-4d58a594797b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b3d39512-98cd-454c-8100-4d58a594797b" (UID: "b3d39512-98cd-454c-8100-4d58a594797b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:44:59 crc kubenswrapper[4923]: I0311 13:44:59.645231 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xd47k\" (UniqueName: \"kubernetes.io/projected/b3d39512-98cd-454c-8100-4d58a594797b-kube-api-access-xd47k\") on node \"crc\" DevicePath \"\"" Mar 11 13:44:59 crc kubenswrapper[4923]: I0311 13:44:59.645267 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3d39512-98cd-454c-8100-4d58a594797b-utilities\") on node \"crc\" DevicePath \"\"" Mar 11 13:44:59 crc kubenswrapper[4923]: I0311 13:44:59.645277 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3d39512-98cd-454c-8100-4d58a594797b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 11 13:44:59 crc kubenswrapper[4923]: I0311 13:44:59.983127 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9vqtj" event={"ID":"b3d39512-98cd-454c-8100-4d58a594797b","Type":"ContainerDied","Data":"e243b858aa912a7d77015009663d61725db6855c6fc527fa12c37ef14b93888a"} Mar 11 13:44:59 crc kubenswrapper[4923]: I0311 13:44:59.983201 4923 scope.go:117] "RemoveContainer" containerID="423a991a4319b8501e3cb40adafe5dd62242132de34d1fab19ac2b3b1db77a86" Mar 11 13:44:59 crc kubenswrapper[4923]: I0311 13:44:59.983285 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9vqtj" Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.023949 4923 scope.go:117] "RemoveContainer" containerID="52369a91e56a0741363a176fa1fd5114f82a9a44636b7498c193a0fb10efe225" Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.047547 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9vqtj"] Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.053696 4923 scope.go:117] "RemoveContainer" containerID="6ab41d0d442faa62ff70a87de2dc3417f01d6d13c0b7b1e3ea49b4407061ff36" Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.059729 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9vqtj"] Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.148965 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29553945-tz8jb"] Mar 11 13:45:00 crc kubenswrapper[4923]: E0311 13:45:00.149415 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d39512-98cd-454c-8100-4d58a594797b" containerName="extract-utilities" Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.149443 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d39512-98cd-454c-8100-4d58a594797b" containerName="extract-utilities" Mar 11 13:45:00 crc kubenswrapper[4923]: E0311 13:45:00.149469 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d39512-98cd-454c-8100-4d58a594797b" containerName="extract-content" Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.149481 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d39512-98cd-454c-8100-4d58a594797b" containerName="extract-content" Mar 11 13:45:00 crc kubenswrapper[4923]: E0311 13:45:00.149505 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3d39512-98cd-454c-8100-4d58a594797b" containerName="registry-server" Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.149519 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3d39512-98cd-454c-8100-4d58a594797b" containerName="registry-server" Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.149715 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3d39512-98cd-454c-8100-4d58a594797b" containerName="registry-server" Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.150447 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29553945-tz8jb" Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.152832 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.153117 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.155276 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29553945-tz8jb"] Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.254248 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/56355c07-0c97-45a2-83dc-2aa197bbc3f3-config-volume\") pod \"collect-profiles-29553945-tz8jb\" (UID: \"56355c07-0c97-45a2-83dc-2aa197bbc3f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553945-tz8jb" Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.254360 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smll4\" (UniqueName: \"kubernetes.io/projected/56355c07-0c97-45a2-83dc-2aa197bbc3f3-kube-api-access-smll4\") pod \"collect-profiles-29553945-tz8jb\" (UID: \"56355c07-0c97-45a2-83dc-2aa197bbc3f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553945-tz8jb" Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.254687 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/56355c07-0c97-45a2-83dc-2aa197bbc3f3-secret-volume\") pod \"collect-profiles-29553945-tz8jb\" (UID: \"56355c07-0c97-45a2-83dc-2aa197bbc3f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553945-tz8jb" Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.356597 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/56355c07-0c97-45a2-83dc-2aa197bbc3f3-config-volume\") pod \"collect-profiles-29553945-tz8jb\" (UID: \"56355c07-0c97-45a2-83dc-2aa197bbc3f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553945-tz8jb" Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.357000 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smll4\" (UniqueName: \"kubernetes.io/projected/56355c07-0c97-45a2-83dc-2aa197bbc3f3-kube-api-access-smll4\") pod \"collect-profiles-29553945-tz8jb\" (UID: \"56355c07-0c97-45a2-83dc-2aa197bbc3f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553945-tz8jb" Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.357113 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/56355c07-0c97-45a2-83dc-2aa197bbc3f3-secret-volume\") pod \"collect-profiles-29553945-tz8jb\" (UID: \"56355c07-0c97-45a2-83dc-2aa197bbc3f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553945-tz8jb" Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.358514 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/56355c07-0c97-45a2-83dc-2aa197bbc3f3-config-volume\") pod \"collect-profiles-29553945-tz8jb\" (UID: \"56355c07-0c97-45a2-83dc-2aa197bbc3f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553945-tz8jb" Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.363899 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/56355c07-0c97-45a2-83dc-2aa197bbc3f3-secret-volume\") pod \"collect-profiles-29553945-tz8jb\" (UID: \"56355c07-0c97-45a2-83dc-2aa197bbc3f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553945-tz8jb" Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.381541 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smll4\" (UniqueName: \"kubernetes.io/projected/56355c07-0c97-45a2-83dc-2aa197bbc3f3-kube-api-access-smll4\") pod \"collect-profiles-29553945-tz8jb\" (UID: \"56355c07-0c97-45a2-83dc-2aa197bbc3f3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29553945-tz8jb" Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.471821 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29553945-tz8jb" Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.715291 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29553945-tz8jb"] Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.992536 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29553945-tz8jb" event={"ID":"56355c07-0c97-45a2-83dc-2aa197bbc3f3","Type":"ContainerStarted","Data":"7833cf5dcae5870592933114b44fa78c620f835db14409e70913880f35e090fc"} Mar 11 13:45:00 crc kubenswrapper[4923]: I0311 13:45:00.992583 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29553945-tz8jb" event={"ID":"56355c07-0c97-45a2-83dc-2aa197bbc3f3","Type":"ContainerStarted","Data":"eab1c708c937de2b68277f3503db7b71dd68c91b5beb324ef333c3ba09ec5250"} Mar 11 13:45:01 crc kubenswrapper[4923]: I0311 13:45:01.022735 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29553945-tz8jb" podStartSLOduration=1.022712577 podStartE2EDuration="1.022712577s" podCreationTimestamp="2026-03-11 13:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-11 13:45:01.01441195 +0000 UTC m=+1348.696313664" watchObservedRunningTime="2026-03-11 13:45:01.022712577 +0000 UTC m=+1348.704614321" Mar 11 13:45:01 crc kubenswrapper[4923]: I0311 13:45:01.039959 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3d39512-98cd-454c-8100-4d58a594797b" path="/var/lib/kubelet/pods/b3d39512-98cd-454c-8100-4d58a594797b/volumes" Mar 11 13:45:02 crc kubenswrapper[4923]: I0311 13:45:02.002414 4923 generic.go:334] "Generic (PLEG): container finished" podID="56355c07-0c97-45a2-83dc-2aa197bbc3f3" containerID="7833cf5dcae5870592933114b44fa78c620f835db14409e70913880f35e090fc" exitCode=0 Mar 11 13:45:02 crc kubenswrapper[4923]: I0311 13:45:02.002472 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29553945-tz8jb" event={"ID":"56355c07-0c97-45a2-83dc-2aa197bbc3f3","Type":"ContainerDied","Data":"7833cf5dcae5870592933114b44fa78c620f835db14409e70913880f35e090fc"} Mar 11 13:45:03 crc kubenswrapper[4923]: I0311 13:45:03.250768 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29553945-tz8jb" Mar 11 13:45:03 crc kubenswrapper[4923]: I0311 13:45:03.307302 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/56355c07-0c97-45a2-83dc-2aa197bbc3f3-secret-volume\") pod \"56355c07-0c97-45a2-83dc-2aa197bbc3f3\" (UID: \"56355c07-0c97-45a2-83dc-2aa197bbc3f3\") " Mar 11 13:45:03 crc kubenswrapper[4923]: I0311 13:45:03.307599 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/56355c07-0c97-45a2-83dc-2aa197bbc3f3-config-volume\") pod \"56355c07-0c97-45a2-83dc-2aa197bbc3f3\" (UID: \"56355c07-0c97-45a2-83dc-2aa197bbc3f3\") " Mar 11 13:45:03 crc kubenswrapper[4923]: I0311 13:45:03.307656 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smll4\" (UniqueName: \"kubernetes.io/projected/56355c07-0c97-45a2-83dc-2aa197bbc3f3-kube-api-access-smll4\") pod \"56355c07-0c97-45a2-83dc-2aa197bbc3f3\" (UID: \"56355c07-0c97-45a2-83dc-2aa197bbc3f3\") " Mar 11 13:45:03 crc kubenswrapper[4923]: I0311 13:45:03.308279 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56355c07-0c97-45a2-83dc-2aa197bbc3f3-config-volume" (OuterVolumeSpecName: "config-volume") pod "56355c07-0c97-45a2-83dc-2aa197bbc3f3" (UID: "56355c07-0c97-45a2-83dc-2aa197bbc3f3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 11 13:45:03 crc kubenswrapper[4923]: I0311 13:45:03.313052 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56355c07-0c97-45a2-83dc-2aa197bbc3f3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "56355c07-0c97-45a2-83dc-2aa197bbc3f3" (UID: "56355c07-0c97-45a2-83dc-2aa197bbc3f3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 11 13:45:03 crc kubenswrapper[4923]: I0311 13:45:03.313086 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56355c07-0c97-45a2-83dc-2aa197bbc3f3-kube-api-access-smll4" (OuterVolumeSpecName: "kube-api-access-smll4") pod "56355c07-0c97-45a2-83dc-2aa197bbc3f3" (UID: "56355c07-0c97-45a2-83dc-2aa197bbc3f3"). InnerVolumeSpecName "kube-api-access-smll4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:45:03 crc kubenswrapper[4923]: I0311 13:45:03.409140 4923 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/56355c07-0c97-45a2-83dc-2aa197bbc3f3-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 11 13:45:03 crc kubenswrapper[4923]: I0311 13:45:03.409182 4923 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/56355c07-0c97-45a2-83dc-2aa197bbc3f3-config-volume\") on node \"crc\" DevicePath \"\"" Mar 11 13:45:03 crc kubenswrapper[4923]: I0311 13:45:03.409196 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smll4\" (UniqueName: \"kubernetes.io/projected/56355c07-0c97-45a2-83dc-2aa197bbc3f3-kube-api-access-smll4\") on node \"crc\" DevicePath \"\"" Mar 11 13:45:04 crc kubenswrapper[4923]: I0311 13:45:04.017673 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29553945-tz8jb" event={"ID":"56355c07-0c97-45a2-83dc-2aa197bbc3f3","Type":"ContainerDied","Data":"eab1c708c937de2b68277f3503db7b71dd68c91b5beb324ef333c3ba09ec5250"} Mar 11 13:45:04 crc kubenswrapper[4923]: I0311 13:45:04.017715 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eab1c708c937de2b68277f3503db7b71dd68c91b5beb324ef333c3ba09ec5250" Mar 11 13:45:04 crc kubenswrapper[4923]: I0311 13:45:04.017764 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29553945-tz8jb" Mar 11 13:45:11 crc kubenswrapper[4923]: I0311 13:45:11.932320 4923 patch_prober.go:28] interesting pod/machine-config-daemon-fsfnw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 11 13:45:11 crc kubenswrapper[4923]: I0311 13:45:11.932706 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 11 13:45:11 crc kubenswrapper[4923]: I0311 13:45:11.932772 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" Mar 11 13:45:11 crc kubenswrapper[4923]: I0311 13:45:11.933620 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3d359e768a3b56caad3d49994ffb4f7bcb69d87648b877548ba08cfe716e455c"} pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 11 13:45:11 crc kubenswrapper[4923]: I0311 13:45:11.933699 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" containerID="cri-o://3d359e768a3b56caad3d49994ffb4f7bcb69d87648b877548ba08cfe716e455c" gracePeriod=600 Mar 11 13:45:12 crc kubenswrapper[4923]: I0311 13:45:12.076205 4923 generic.go:334] "Generic (PLEG): container finished" podID="356de301-ded3-4112-8406-1a28e44c2082" containerID="3d359e768a3b56caad3d49994ffb4f7bcb69d87648b877548ba08cfe716e455c" exitCode=0 Mar 11 13:45:12 crc kubenswrapper[4923]: I0311 13:45:12.076296 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" event={"ID":"356de301-ded3-4112-8406-1a28e44c2082","Type":"ContainerDied","Data":"3d359e768a3b56caad3d49994ffb4f7bcb69d87648b877548ba08cfe716e455c"} Mar 11 13:45:12 crc kubenswrapper[4923]: I0311 13:45:12.076670 4923 scope.go:117] "RemoveContainer" containerID="ead0c2f701ab9961c880f0090716c64783b2bee2562b04e2163c46d120ab88f4" Mar 11 13:45:12 crc kubenswrapper[4923]: I0311 13:45:12.779013 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-9cjnl_00adaa61-ef98-4a06-9f60-4bd990cce287/control-plane-machine-set-operator/0.log" Mar 11 13:45:12 crc kubenswrapper[4923]: I0311 13:45:12.877321 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-c7dks_9a054bc8-9c59-47f8-8b37-7eb59a5331a0/kube-rbac-proxy/0.log" Mar 11 13:45:12 crc kubenswrapper[4923]: I0311 13:45:12.967149 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-c7dks_9a054bc8-9c59-47f8-8b37-7eb59a5331a0/machine-api-operator/0.log" Mar 11 13:45:13 crc kubenswrapper[4923]: I0311 13:45:13.084904 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" event={"ID":"356de301-ded3-4112-8406-1a28e44c2082","Type":"ContainerStarted","Data":"c22d0f7656c6b989475681637e5cd2d3b3a534ac5fa3c03b27449f2838f08907"} Mar 11 13:45:43 crc kubenswrapper[4923]: I0311 13:45:43.018917 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-47vwc_475686c9-3c06-456e-b570-5b3774a834d9/kube-rbac-proxy/0.log" Mar 11 13:45:43 crc kubenswrapper[4923]: I0311 13:45:43.066108 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-47vwc_475686c9-3c06-456e-b570-5b3774a834d9/controller/0.log" Mar 11 13:45:43 crc kubenswrapper[4923]: I0311 13:45:43.206014 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9f8w7_21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0/cp-frr-files/0.log" Mar 11 13:45:43 crc kubenswrapper[4923]: I0311 13:45:43.385688 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9f8w7_21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0/cp-reloader/0.log" Mar 11 13:45:43 crc kubenswrapper[4923]: I0311 13:45:43.421231 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9f8w7_21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0/cp-frr-files/0.log" Mar 11 13:45:43 crc kubenswrapper[4923]: I0311 13:45:43.445840 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9f8w7_21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0/cp-metrics/0.log" Mar 11 13:45:43 crc kubenswrapper[4923]: I0311 13:45:43.448518 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9f8w7_21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0/cp-reloader/0.log" Mar 11 13:45:43 crc kubenswrapper[4923]: I0311 13:45:43.584504 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9f8w7_21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0/cp-reloader/0.log" Mar 11 13:45:43 crc kubenswrapper[4923]: I0311 13:45:43.605753 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9f8w7_21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0/cp-frr-files/0.log" Mar 11 13:45:43 crc kubenswrapper[4923]: I0311 13:45:43.635103 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9f8w7_21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0/cp-metrics/0.log" Mar 11 13:45:43 crc kubenswrapper[4923]: I0311 13:45:43.639497 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9f8w7_21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0/cp-metrics/0.log" Mar 11 13:45:43 crc kubenswrapper[4923]: I0311 13:45:43.789670 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9f8w7_21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0/cp-frr-files/0.log" Mar 11 13:45:43 crc kubenswrapper[4923]: I0311 13:45:43.795557 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9f8w7_21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0/controller/0.log" Mar 11 13:45:43 crc kubenswrapper[4923]: I0311 13:45:43.814962 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9f8w7_21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0/cp-reloader/0.log" Mar 11 13:45:43 crc kubenswrapper[4923]: I0311 13:45:43.816353 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9f8w7_21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0/cp-metrics/0.log" Mar 11 13:45:43 crc kubenswrapper[4923]: I0311 13:45:43.965825 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9f8w7_21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0/frr-metrics/0.log" Mar 11 13:45:43 crc kubenswrapper[4923]: I0311 13:45:43.973718 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9f8w7_21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0/kube-rbac-proxy/0.log" Mar 11 13:45:43 crc kubenswrapper[4923]: I0311 13:45:43.999066 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9f8w7_21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0/kube-rbac-proxy-frr/0.log" Mar 11 13:45:44 crc kubenswrapper[4923]: I0311 13:45:44.222407 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-zwkrg_0085e43b-b941-4b8b-b0c3-3ca1cdd68c58/frr-k8s-webhook-server/0.log" Mar 11 13:45:44 crc kubenswrapper[4923]: I0311 13:45:44.233502 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9f8w7_21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0/reloader/0.log" Mar 11 13:45:44 crc kubenswrapper[4923]: I0311 13:45:44.269988 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-9f8w7_21b2ab18-baa1-4fdb-8ae2-f8ab2a877ed0/frr/0.log" Mar 11 13:45:44 crc kubenswrapper[4923]: I0311 13:45:44.435013 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-8547b8bf69-qqn9m_4be74972-4fdd-48b0-bb91-462c20ecec81/manager/0.log" Mar 11 13:45:44 crc kubenswrapper[4923]: I0311 13:45:44.447301 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-5865df787d-xplvf_a56ae696-0c98-464b-a2de-75204772866b/webhook-server/0.log" Mar 11 13:45:44 crc kubenswrapper[4923]: I0311 13:45:44.624947 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-rrmgz_a4d6dd0e-4763-486d-a6a6-8ce519149aaa/kube-rbac-proxy/0.log" Mar 11 13:45:44 crc kubenswrapper[4923]: I0311 13:45:44.670693 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-rrmgz_a4d6dd0e-4763-486d-a6a6-8ce519149aaa/speaker/0.log" Mar 11 13:46:00 crc kubenswrapper[4923]: I0311 13:46:00.144473 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29553946-rpbrd"] Mar 11 13:46:00 crc kubenswrapper[4923]: E0311 13:46:00.145505 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56355c07-0c97-45a2-83dc-2aa197bbc3f3" containerName="collect-profiles" Mar 11 13:46:00 crc kubenswrapper[4923]: I0311 13:46:00.145519 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="56355c07-0c97-45a2-83dc-2aa197bbc3f3" containerName="collect-profiles" Mar 11 13:46:00 crc kubenswrapper[4923]: I0311 13:46:00.145770 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="56355c07-0c97-45a2-83dc-2aa197bbc3f3" containerName="collect-profiles" Mar 11 13:46:00 crc kubenswrapper[4923]: I0311 13:46:00.146776 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553946-rpbrd" Mar 11 13:46:00 crc kubenswrapper[4923]: I0311 13:46:00.152732 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 11 13:46:00 crc kubenswrapper[4923]: I0311 13:46:00.152871 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-nxck7" Mar 11 13:46:00 crc kubenswrapper[4923]: I0311 13:46:00.152914 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 11 13:46:00 crc kubenswrapper[4923]: I0311 13:46:00.154298 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/cinder-kuttl-tests_keystone-6b4558db79-pn5px_bda78896-5134-4b9b-917d-40801488fcaa/keystone-api/0.log" Mar 11 13:46:00 crc kubenswrapper[4923]: I0311 13:46:00.155057 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553946-rpbrd"] Mar 11 13:46:00 crc kubenswrapper[4923]: I0311 13:46:00.268426 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdzvw\" (UniqueName: \"kubernetes.io/projected/a87624ea-d4e7-4558-bec6-6db67ea015dd-kube-api-access-fdzvw\") pod \"auto-csr-approver-29553946-rpbrd\" (UID: \"a87624ea-d4e7-4558-bec6-6db67ea015dd\") " pod="openshift-infra/auto-csr-approver-29553946-rpbrd" Mar 11 13:46:00 crc kubenswrapper[4923]: I0311 13:46:00.274736 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/cinder-kuttl-tests_keystone-bootstrap-nk5mc_7ad98c9d-3330-47ad-bf1e-a208ad0401ce/keystone-bootstrap/0.log" Mar 11 13:46:00 crc kubenswrapper[4923]: I0311 13:46:00.295793 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/cinder-kuttl-tests_keystone-db-create-clm8k_60b9b9dd-cd32-48bb-9995-36bf024f9087/mariadb-database-create/0.log" Mar 11 13:46:00 crc kubenswrapper[4923]: I0311 13:46:00.369779 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdzvw\" (UniqueName: \"kubernetes.io/projected/a87624ea-d4e7-4558-bec6-6db67ea015dd-kube-api-access-fdzvw\") pod \"auto-csr-approver-29553946-rpbrd\" (UID: \"a87624ea-d4e7-4558-bec6-6db67ea015dd\") " pod="openshift-infra/auto-csr-approver-29553946-rpbrd" Mar 11 13:46:00 crc kubenswrapper[4923]: I0311 13:46:00.395183 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdzvw\" (UniqueName: \"kubernetes.io/projected/a87624ea-d4e7-4558-bec6-6db67ea015dd-kube-api-access-fdzvw\") pod \"auto-csr-approver-29553946-rpbrd\" (UID: \"a87624ea-d4e7-4558-bec6-6db67ea015dd\") " pod="openshift-infra/auto-csr-approver-29553946-rpbrd" Mar 11 13:46:00 crc kubenswrapper[4923]: I0311 13:46:00.466878 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553946-rpbrd" Mar 11 13:46:00 crc kubenswrapper[4923]: I0311 13:46:00.632508 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/cinder-kuttl-tests_keystone-f524-account-create-update-cdc7k_19fc67d1-7c03-4a55-8baa-362e5a8a2344/mariadb-account-create-update/0.log" Mar 11 13:46:00 crc kubenswrapper[4923]: I0311 13:46:00.660747 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/cinder-kuttl-tests_keystone-db-sync-cjv6x_eae8474f-7523-4ee9-8f02-7be328688043/keystone-db-sync/0.log" Mar 11 13:46:00 crc kubenswrapper[4923]: I0311 13:46:00.853761 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/cinder-kuttl-tests_openstack-galera-0_229f08c8-d2cf-46a7-b933-cb5223b654ff/mysql-bootstrap/0.log" Mar 11 13:46:00 crc kubenswrapper[4923]: I0311 13:46:00.891817 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553946-rpbrd"] Mar 11 13:46:01 crc kubenswrapper[4923]: I0311 13:46:01.033938 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/cinder-kuttl-tests_openstack-galera-0_229f08c8-d2cf-46a7-b933-cb5223b654ff/mysql-bootstrap/0.log" Mar 11 13:46:01 crc kubenswrapper[4923]: I0311 13:46:01.067730 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/cinder-kuttl-tests_memcached-0_6ccfe8a9-edc8-41a2-af22-d55181b3dba1/memcached/0.log" Mar 11 13:46:01 crc kubenswrapper[4923]: I0311 13:46:01.106357 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/cinder-kuttl-tests_openstack-galera-0_229f08c8-d2cf-46a7-b933-cb5223b654ff/galera/0.log" Mar 11 13:46:01 crc kubenswrapper[4923]: I0311 13:46:01.208225 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/cinder-kuttl-tests_openstack-galera-1_5b12f995-e853-4d67-a013-beb96a7ac30b/mysql-bootstrap/0.log" Mar 11 13:46:01 crc kubenswrapper[4923]: I0311 13:46:01.370218 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/cinder-kuttl-tests_openstack-galera-1_5b12f995-e853-4d67-a013-beb96a7ac30b/galera/0.log" Mar 11 13:46:01 crc kubenswrapper[4923]: I0311 13:46:01.405120 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/cinder-kuttl-tests_openstack-galera-2_57cc1bb8-4fe5-42af-b6b8-672d0baf4b06/mysql-bootstrap/0.log" Mar 11 13:46:01 crc kubenswrapper[4923]: I0311 13:46:01.405142 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/cinder-kuttl-tests_openstack-galera-1_5b12f995-e853-4d67-a013-beb96a7ac30b/mysql-bootstrap/0.log" Mar 11 13:46:01 crc kubenswrapper[4923]: I0311 13:46:01.424678 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553946-rpbrd" event={"ID":"a87624ea-d4e7-4558-bec6-6db67ea015dd","Type":"ContainerStarted","Data":"7447c2b2fbdc9f2da7118d0cffb61ddf603c3c2d3fa5c950cf698ca6d2239867"} Mar 11 13:46:01 crc kubenswrapper[4923]: I0311 13:46:01.580961 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/cinder-kuttl-tests_openstack-galera-2_57cc1bb8-4fe5-42af-b6b8-672d0baf4b06/mysql-bootstrap/0.log" Mar 11 13:46:01 crc kubenswrapper[4923]: I0311 13:46:01.594055 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/cinder-kuttl-tests_openstack-galera-2_57cc1bb8-4fe5-42af-b6b8-672d0baf4b06/galera/0.log" Mar 11 13:46:01 crc kubenswrapper[4923]: I0311 13:46:01.657374 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/cinder-kuttl-tests_rabbitmq-server-0_50119944-78f3-4b19-b19b-1bc87e93e3fc/setup-container/0.log" Mar 11 13:46:01 crc kubenswrapper[4923]: I0311 13:46:01.780951 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/cinder-kuttl-tests_rabbitmq-server-0_50119944-78f3-4b19-b19b-1bc87e93e3fc/setup-container/0.log" Mar 11 13:46:01 crc kubenswrapper[4923]: I0311 13:46:01.804847 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/cinder-kuttl-tests_rabbitmq-server-0_50119944-78f3-4b19-b19b-1bc87e93e3fc/rabbitmq/0.log" Mar 11 13:46:01 crc kubenswrapper[4923]: I0311 13:46:01.838774 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/cinder-kuttl-tests_root-account-create-update-hps8z_1da28d4d-80b8-4a02-8009-0b6032044448/mariadb-account-create-update/0.log" Mar 11 13:46:02 crc kubenswrapper[4923]: I0311 13:46:02.431744 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553946-rpbrd" event={"ID":"a87624ea-d4e7-4558-bec6-6db67ea015dd","Type":"ContainerStarted","Data":"8b6846ef7da7ff7445e5369e0c5e0ba5d960e05c39cfe851311f7ccaee0265de"} Mar 11 13:46:03 crc kubenswrapper[4923]: I0311 13:46:03.442475 4923 generic.go:334] "Generic (PLEG): container finished" podID="a87624ea-d4e7-4558-bec6-6db67ea015dd" containerID="8b6846ef7da7ff7445e5369e0c5e0ba5d960e05c39cfe851311f7ccaee0265de" exitCode=0 Mar 11 13:46:03 crc kubenswrapper[4923]: I0311 13:46:03.442520 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553946-rpbrd" event={"ID":"a87624ea-d4e7-4558-bec6-6db67ea015dd","Type":"ContainerDied","Data":"8b6846ef7da7ff7445e5369e0c5e0ba5d960e05c39cfe851311f7ccaee0265de"} Mar 11 13:46:04 crc kubenswrapper[4923]: I0311 13:46:04.715496 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553946-rpbrd" Mar 11 13:46:04 crc kubenswrapper[4923]: I0311 13:46:04.831561 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdzvw\" (UniqueName: \"kubernetes.io/projected/a87624ea-d4e7-4558-bec6-6db67ea015dd-kube-api-access-fdzvw\") pod \"a87624ea-d4e7-4558-bec6-6db67ea015dd\" (UID: \"a87624ea-d4e7-4558-bec6-6db67ea015dd\") " Mar 11 13:46:04 crc kubenswrapper[4923]: I0311 13:46:04.841677 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a87624ea-d4e7-4558-bec6-6db67ea015dd-kube-api-access-fdzvw" (OuterVolumeSpecName: "kube-api-access-fdzvw") pod "a87624ea-d4e7-4558-bec6-6db67ea015dd" (UID: "a87624ea-d4e7-4558-bec6-6db67ea015dd"). InnerVolumeSpecName "kube-api-access-fdzvw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:46:04 crc kubenswrapper[4923]: I0311 13:46:04.933544 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdzvw\" (UniqueName: \"kubernetes.io/projected/a87624ea-d4e7-4558-bec6-6db67ea015dd-kube-api-access-fdzvw\") on node \"crc\" DevicePath \"\"" Mar 11 13:46:05 crc kubenswrapper[4923]: I0311 13:46:05.459696 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553946-rpbrd" event={"ID":"a87624ea-d4e7-4558-bec6-6db67ea015dd","Type":"ContainerDied","Data":"7447c2b2fbdc9f2da7118d0cffb61ddf603c3c2d3fa5c950cf698ca6d2239867"} Mar 11 13:46:05 crc kubenswrapper[4923]: I0311 13:46:05.460034 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7447c2b2fbdc9f2da7118d0cffb61ddf603c3c2d3fa5c950cf698ca6d2239867" Mar 11 13:46:05 crc kubenswrapper[4923]: I0311 13:46:05.459772 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553946-rpbrd" Mar 11 13:46:05 crc kubenswrapper[4923]: I0311 13:46:05.513634 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29553940-wj5ml"] Mar 11 13:46:05 crc kubenswrapper[4923]: I0311 13:46:05.519913 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29553940-wj5ml"] Mar 11 13:46:07 crc kubenswrapper[4923]: I0311 13:46:07.037226 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13c70bc8-3f75-401b-8a96-d3d044e39672" path="/var/lib/kubelet/pods/13c70bc8-3f75-401b-8a96-d3d044e39672/volumes" Mar 11 13:46:15 crc kubenswrapper[4923]: I0311 13:46:15.670021 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww_45210d52-d52e-4809-b571-ca076d6e9d84/util/0.log" Mar 11 13:46:15 crc kubenswrapper[4923]: I0311 13:46:15.809562 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww_45210d52-d52e-4809-b571-ca076d6e9d84/pull/0.log" Mar 11 13:46:15 crc kubenswrapper[4923]: I0311 13:46:15.816613 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww_45210d52-d52e-4809-b571-ca076d6e9d84/pull/0.log" Mar 11 13:46:15 crc kubenswrapper[4923]: I0311 13:46:15.843650 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww_45210d52-d52e-4809-b571-ca076d6e9d84/util/0.log" Mar 11 13:46:16 crc kubenswrapper[4923]: I0311 13:46:16.004033 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww_45210d52-d52e-4809-b571-ca076d6e9d84/util/0.log" Mar 11 13:46:16 crc kubenswrapper[4923]: I0311 13:46:16.015902 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww_45210d52-d52e-4809-b571-ca076d6e9d84/pull/0.log" Mar 11 13:46:16 crc kubenswrapper[4923]: I0311 13:46:16.032046 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c19w5ww_45210d52-d52e-4809-b571-ca076d6e9d84/extract/0.log" Mar 11 13:46:16 crc kubenswrapper[4923]: I0311 13:46:16.178997 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zsbfg_38a8794f-af70-4f93-a892-13e5e6ccae32/extract-utilities/0.log" Mar 11 13:46:16 crc kubenswrapper[4923]: I0311 13:46:16.318414 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zsbfg_38a8794f-af70-4f93-a892-13e5e6ccae32/extract-content/0.log" Mar 11 13:46:16 crc kubenswrapper[4923]: I0311 13:46:16.319400 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zsbfg_38a8794f-af70-4f93-a892-13e5e6ccae32/extract-utilities/0.log" Mar 11 13:46:16 crc kubenswrapper[4923]: I0311 13:46:16.351170 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zsbfg_38a8794f-af70-4f93-a892-13e5e6ccae32/extract-content/0.log" Mar 11 13:46:16 crc kubenswrapper[4923]: I0311 13:46:16.522732 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zsbfg_38a8794f-af70-4f93-a892-13e5e6ccae32/extract-utilities/0.log" Mar 11 13:46:16 crc kubenswrapper[4923]: I0311 13:46:16.570999 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zsbfg_38a8794f-af70-4f93-a892-13e5e6ccae32/extract-content/0.log" Mar 11 13:46:16 crc kubenswrapper[4923]: I0311 13:46:16.710211 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j8cwt_5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9/extract-utilities/0.log" Mar 11 13:46:16 crc kubenswrapper[4923]: I0311 13:46:16.805887 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zsbfg_38a8794f-af70-4f93-a892-13e5e6ccae32/registry-server/0.log" Mar 11 13:46:16 crc kubenswrapper[4923]: I0311 13:46:16.918317 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j8cwt_5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9/extract-content/0.log" Mar 11 13:46:16 crc kubenswrapper[4923]: I0311 13:46:16.930792 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j8cwt_5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9/extract-content/0.log" Mar 11 13:46:16 crc kubenswrapper[4923]: I0311 13:46:16.932086 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j8cwt_5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9/extract-utilities/0.log" Mar 11 13:46:17 crc kubenswrapper[4923]: I0311 13:46:17.129429 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j8cwt_5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9/extract-utilities/0.log" Mar 11 13:46:17 crc kubenswrapper[4923]: I0311 13:46:17.153391 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j8cwt_5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9/extract-content/0.log" Mar 11 13:46:17 crc kubenswrapper[4923]: I0311 13:46:17.399440 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-j8cwt_5986b4bc-9bbc-43a2-ab9b-2bcf145c7fb9/registry-server/0.log" Mar 11 13:46:17 crc kubenswrapper[4923]: I0311 13:46:17.414284 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-z7n2v_85ba46fa-fad0-421a-8b0a-e83a8e417079/marketplace-operator/0.log" Mar 11 13:46:17 crc kubenswrapper[4923]: I0311 13:46:17.467022 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x98sf_fe7957fa-3f27-4325-8e3b-5593c750bb6b/extract-utilities/0.log" Mar 11 13:46:17 crc kubenswrapper[4923]: I0311 13:46:17.641015 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x98sf_fe7957fa-3f27-4325-8e3b-5593c750bb6b/extract-utilities/0.log" Mar 11 13:46:17 crc kubenswrapper[4923]: I0311 13:46:17.643180 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x98sf_fe7957fa-3f27-4325-8e3b-5593c750bb6b/extract-content/0.log" Mar 11 13:46:17 crc kubenswrapper[4923]: I0311 13:46:17.645603 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x98sf_fe7957fa-3f27-4325-8e3b-5593c750bb6b/extract-content/0.log" Mar 11 13:46:17 crc kubenswrapper[4923]: I0311 13:46:17.846430 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x98sf_fe7957fa-3f27-4325-8e3b-5593c750bb6b/extract-utilities/0.log" Mar 11 13:46:17 crc kubenswrapper[4923]: I0311 13:46:17.852168 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x98sf_fe7957fa-3f27-4325-8e3b-5593c750bb6b/extract-content/0.log" Mar 11 13:46:17 crc kubenswrapper[4923]: I0311 13:46:17.937388 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-x98sf_fe7957fa-3f27-4325-8e3b-5593c750bb6b/registry-server/0.log" Mar 11 13:46:18 crc kubenswrapper[4923]: I0311 13:46:18.058477 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-48q6k_99a57330-6112-4e39-aec6-3f62ff4a56ea/extract-utilities/0.log" Mar 11 13:46:18 crc kubenswrapper[4923]: I0311 13:46:18.221525 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-48q6k_99a57330-6112-4e39-aec6-3f62ff4a56ea/extract-utilities/0.log" Mar 11 13:46:18 crc kubenswrapper[4923]: I0311 13:46:18.243421 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-48q6k_99a57330-6112-4e39-aec6-3f62ff4a56ea/extract-content/0.log" Mar 11 13:46:18 crc kubenswrapper[4923]: I0311 13:46:18.265712 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-48q6k_99a57330-6112-4e39-aec6-3f62ff4a56ea/extract-content/0.log" Mar 11 13:46:18 crc kubenswrapper[4923]: I0311 13:46:18.391771 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-48q6k_99a57330-6112-4e39-aec6-3f62ff4a56ea/extract-utilities/0.log" Mar 11 13:46:18 crc kubenswrapper[4923]: I0311 13:46:18.433302 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-48q6k_99a57330-6112-4e39-aec6-3f62ff4a56ea/extract-content/0.log" Mar 11 13:46:18 crc kubenswrapper[4923]: I0311 13:46:18.695676 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-48q6k_99a57330-6112-4e39-aec6-3f62ff4a56ea/registry-server/0.log" Mar 11 13:46:34 crc kubenswrapper[4923]: I0311 13:46:34.348183 4923 scope.go:117] "RemoveContainer" containerID="e437e1c2f3c3617a128fbb5b3bb5b67e8a7d4caa69c51814e865f73459311004" Mar 11 13:46:34 crc kubenswrapper[4923]: I0311 13:46:34.382787 4923 scope.go:117] "RemoveContainer" containerID="fb6ab8fb89f37fb010e2ca58da098bf09d4aeb72cf5ee3378de079317faf467c" Mar 11 13:46:34 crc kubenswrapper[4923]: I0311 13:46:34.431766 4923 scope.go:117] "RemoveContainer" containerID="fca8ac0b773cafac5f5ef9ba283429f0847b0ce13ccb570af77cb877754851e2" Mar 11 13:46:34 crc kubenswrapper[4923]: I0311 13:46:34.451928 4923 scope.go:117] "RemoveContainer" containerID="304bb4d25b0ee4a7e44d56d37696f9762d3859438b560ce2b1936ba1a46dd358" Mar 11 13:47:27 crc kubenswrapper[4923]: I0311 13:47:27.067506 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/root-account-create-update-hps8z"] Mar 11 13:47:27 crc kubenswrapper[4923]: I0311 13:47:27.074618 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/root-account-create-update-hps8z"] Mar 11 13:47:29 crc kubenswrapper[4923]: I0311 13:47:29.035539 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1da28d4d-80b8-4a02-8009-0b6032044448" path="/var/lib/kubelet/pods/1da28d4d-80b8-4a02-8009-0b6032044448/volumes" Mar 11 13:47:30 crc kubenswrapper[4923]: I0311 13:47:30.234364 4923 generic.go:334] "Generic (PLEG): container finished" podID="931f29a0-a23b-4191-9abd-19026cb35843" containerID="8ed5b7eb28064e58bae9d9bea8486d9ae035b6b14290d20e70c1106ef634bc8f" exitCode=0 Mar 11 13:47:30 crc kubenswrapper[4923]: I0311 13:47:30.234410 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-h9dp2/must-gather-fn4dg" event={"ID":"931f29a0-a23b-4191-9abd-19026cb35843","Type":"ContainerDied","Data":"8ed5b7eb28064e58bae9d9bea8486d9ae035b6b14290d20e70c1106ef634bc8f"} Mar 11 13:47:30 crc kubenswrapper[4923]: I0311 13:47:30.234918 4923 scope.go:117] "RemoveContainer" containerID="8ed5b7eb28064e58bae9d9bea8486d9ae035b6b14290d20e70c1106ef634bc8f" Mar 11 13:47:31 crc kubenswrapper[4923]: I0311 13:47:31.277457 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-h9dp2_must-gather-fn4dg_931f29a0-a23b-4191-9abd-19026cb35843/gather/0.log" Mar 11 13:47:34 crc kubenswrapper[4923]: I0311 13:47:34.556242 4923 scope.go:117] "RemoveContainer" containerID="c773a542a34bd7b1ca88cfa25860316a4806ff4032287ddf0e84a23098fb6f22" Mar 11 13:47:38 crc kubenswrapper[4923]: I0311 13:47:38.718541 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-h9dp2/must-gather-fn4dg"] Mar 11 13:47:38 crc kubenswrapper[4923]: I0311 13:47:38.719098 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-h9dp2/must-gather-fn4dg" podUID="931f29a0-a23b-4191-9abd-19026cb35843" containerName="copy" containerID="cri-o://ce723073b20c7bb87e20d9abfbe598299afabe97f123e455b9c390f55398155a" gracePeriod=2 Mar 11 13:47:38 crc kubenswrapper[4923]: I0311 13:47:38.730983 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-h9dp2/must-gather-fn4dg"] Mar 11 13:47:39 crc kubenswrapper[4923]: I0311 13:47:39.099662 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-h9dp2_must-gather-fn4dg_931f29a0-a23b-4191-9abd-19026cb35843/copy/0.log" Mar 11 13:47:39 crc kubenswrapper[4923]: I0311 13:47:39.100126 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h9dp2/must-gather-fn4dg" Mar 11 13:47:39 crc kubenswrapper[4923]: I0311 13:47:39.259061 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/931f29a0-a23b-4191-9abd-19026cb35843-must-gather-output\") pod \"931f29a0-a23b-4191-9abd-19026cb35843\" (UID: \"931f29a0-a23b-4191-9abd-19026cb35843\") " Mar 11 13:47:39 crc kubenswrapper[4923]: I0311 13:47:39.259161 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmgkj\" (UniqueName: \"kubernetes.io/projected/931f29a0-a23b-4191-9abd-19026cb35843-kube-api-access-xmgkj\") pod \"931f29a0-a23b-4191-9abd-19026cb35843\" (UID: \"931f29a0-a23b-4191-9abd-19026cb35843\") " Mar 11 13:47:39 crc kubenswrapper[4923]: I0311 13:47:39.265107 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/931f29a0-a23b-4191-9abd-19026cb35843-kube-api-access-xmgkj" (OuterVolumeSpecName: "kube-api-access-xmgkj") pod "931f29a0-a23b-4191-9abd-19026cb35843" (UID: "931f29a0-a23b-4191-9abd-19026cb35843"). InnerVolumeSpecName "kube-api-access-xmgkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:47:39 crc kubenswrapper[4923]: I0311 13:47:39.314669 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-h9dp2_must-gather-fn4dg_931f29a0-a23b-4191-9abd-19026cb35843/copy/0.log" Mar 11 13:47:39 crc kubenswrapper[4923]: I0311 13:47:39.314952 4923 generic.go:334] "Generic (PLEG): container finished" podID="931f29a0-a23b-4191-9abd-19026cb35843" containerID="ce723073b20c7bb87e20d9abfbe598299afabe97f123e455b9c390f55398155a" exitCode=143 Mar 11 13:47:39 crc kubenswrapper[4923]: I0311 13:47:39.314995 4923 scope.go:117] "RemoveContainer" containerID="ce723073b20c7bb87e20d9abfbe598299afabe97f123e455b9c390f55398155a" Mar 11 13:47:39 crc kubenswrapper[4923]: I0311 13:47:39.315090 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-h9dp2/must-gather-fn4dg" Mar 11 13:47:39 crc kubenswrapper[4923]: I0311 13:47:39.332033 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/931f29a0-a23b-4191-9abd-19026cb35843-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "931f29a0-a23b-4191-9abd-19026cb35843" (UID: "931f29a0-a23b-4191-9abd-19026cb35843"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:47:39 crc kubenswrapper[4923]: I0311 13:47:39.343978 4923 scope.go:117] "RemoveContainer" containerID="8ed5b7eb28064e58bae9d9bea8486d9ae035b6b14290d20e70c1106ef634bc8f" Mar 11 13:47:39 crc kubenswrapper[4923]: I0311 13:47:39.361439 4923 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/931f29a0-a23b-4191-9abd-19026cb35843-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 11 13:47:39 crc kubenswrapper[4923]: I0311 13:47:39.361485 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmgkj\" (UniqueName: \"kubernetes.io/projected/931f29a0-a23b-4191-9abd-19026cb35843-kube-api-access-xmgkj\") on node \"crc\" DevicePath \"\"" Mar 11 13:47:39 crc kubenswrapper[4923]: I0311 13:47:39.381628 4923 scope.go:117] "RemoveContainer" containerID="ce723073b20c7bb87e20d9abfbe598299afabe97f123e455b9c390f55398155a" Mar 11 13:47:39 crc kubenswrapper[4923]: E0311 13:47:39.382029 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce723073b20c7bb87e20d9abfbe598299afabe97f123e455b9c390f55398155a\": container with ID starting with ce723073b20c7bb87e20d9abfbe598299afabe97f123e455b9c390f55398155a not found: ID does not exist" containerID="ce723073b20c7bb87e20d9abfbe598299afabe97f123e455b9c390f55398155a" Mar 11 13:47:39 crc kubenswrapper[4923]: I0311 13:47:39.382058 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce723073b20c7bb87e20d9abfbe598299afabe97f123e455b9c390f55398155a"} err="failed to get container status \"ce723073b20c7bb87e20d9abfbe598299afabe97f123e455b9c390f55398155a\": rpc error: code = NotFound desc = could not find container \"ce723073b20c7bb87e20d9abfbe598299afabe97f123e455b9c390f55398155a\": container with ID starting with ce723073b20c7bb87e20d9abfbe598299afabe97f123e455b9c390f55398155a not found: ID does not exist" Mar 11 13:47:39 crc kubenswrapper[4923]: I0311 13:47:39.382079 4923 scope.go:117] "RemoveContainer" containerID="8ed5b7eb28064e58bae9d9bea8486d9ae035b6b14290d20e70c1106ef634bc8f" Mar 11 13:47:39 crc kubenswrapper[4923]: E0311 13:47:39.382300 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ed5b7eb28064e58bae9d9bea8486d9ae035b6b14290d20e70c1106ef634bc8f\": container with ID starting with 8ed5b7eb28064e58bae9d9bea8486d9ae035b6b14290d20e70c1106ef634bc8f not found: ID does not exist" containerID="8ed5b7eb28064e58bae9d9bea8486d9ae035b6b14290d20e70c1106ef634bc8f" Mar 11 13:47:39 crc kubenswrapper[4923]: I0311 13:47:39.382321 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ed5b7eb28064e58bae9d9bea8486d9ae035b6b14290d20e70c1106ef634bc8f"} err="failed to get container status \"8ed5b7eb28064e58bae9d9bea8486d9ae035b6b14290d20e70c1106ef634bc8f\": rpc error: code = NotFound desc = could not find container \"8ed5b7eb28064e58bae9d9bea8486d9ae035b6b14290d20e70c1106ef634bc8f\": container with ID starting with 8ed5b7eb28064e58bae9d9bea8486d9ae035b6b14290d20e70c1106ef634bc8f not found: ID does not exist" Mar 11 13:47:41 crc kubenswrapper[4923]: I0311 13:47:41.039494 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="931f29a0-a23b-4191-9abd-19026cb35843" path="/var/lib/kubelet/pods/931f29a0-a23b-4191-9abd-19026cb35843/volumes" Mar 11 13:47:41 crc kubenswrapper[4923]: I0311 13:47:41.932266 4923 patch_prober.go:28] interesting pod/machine-config-daemon-fsfnw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 11 13:47:41 crc kubenswrapper[4923]: I0311 13:47:41.932408 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 11 13:48:00 crc kubenswrapper[4923]: I0311 13:48:00.155333 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29553948-6trm4"] Mar 11 13:48:00 crc kubenswrapper[4923]: E0311 13:48:00.156560 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="931f29a0-a23b-4191-9abd-19026cb35843" containerName="copy" Mar 11 13:48:00 crc kubenswrapper[4923]: I0311 13:48:00.156587 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="931f29a0-a23b-4191-9abd-19026cb35843" containerName="copy" Mar 11 13:48:00 crc kubenswrapper[4923]: E0311 13:48:00.156602 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a87624ea-d4e7-4558-bec6-6db67ea015dd" containerName="oc" Mar 11 13:48:00 crc kubenswrapper[4923]: I0311 13:48:00.156610 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="a87624ea-d4e7-4558-bec6-6db67ea015dd" containerName="oc" Mar 11 13:48:00 crc kubenswrapper[4923]: E0311 13:48:00.156623 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="931f29a0-a23b-4191-9abd-19026cb35843" containerName="gather" Mar 11 13:48:00 crc kubenswrapper[4923]: I0311 13:48:00.156632 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="931f29a0-a23b-4191-9abd-19026cb35843" containerName="gather" Mar 11 13:48:00 crc kubenswrapper[4923]: I0311 13:48:00.156775 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="a87624ea-d4e7-4558-bec6-6db67ea015dd" containerName="oc" Mar 11 13:48:00 crc kubenswrapper[4923]: I0311 13:48:00.156792 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="931f29a0-a23b-4191-9abd-19026cb35843" containerName="copy" Mar 11 13:48:00 crc kubenswrapper[4923]: I0311 13:48:00.156807 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="931f29a0-a23b-4191-9abd-19026cb35843" containerName="gather" Mar 11 13:48:00 crc kubenswrapper[4923]: I0311 13:48:00.157529 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553948-6trm4" Mar 11 13:48:00 crc kubenswrapper[4923]: I0311 13:48:00.164005 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-nxck7" Mar 11 13:48:00 crc kubenswrapper[4923]: I0311 13:48:00.164057 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 11 13:48:00 crc kubenswrapper[4923]: I0311 13:48:00.164560 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 11 13:48:00 crc kubenswrapper[4923]: I0311 13:48:00.171784 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553948-6trm4"] Mar 11 13:48:00 crc kubenswrapper[4923]: I0311 13:48:00.279164 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bj4z\" (UniqueName: \"kubernetes.io/projected/8d9038fe-7cb9-4093-87b8-ec23182fd21d-kube-api-access-7bj4z\") pod \"auto-csr-approver-29553948-6trm4\" (UID: \"8d9038fe-7cb9-4093-87b8-ec23182fd21d\") " pod="openshift-infra/auto-csr-approver-29553948-6trm4" Mar 11 13:48:00 crc kubenswrapper[4923]: I0311 13:48:00.380291 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bj4z\" (UniqueName: \"kubernetes.io/projected/8d9038fe-7cb9-4093-87b8-ec23182fd21d-kube-api-access-7bj4z\") pod \"auto-csr-approver-29553948-6trm4\" (UID: \"8d9038fe-7cb9-4093-87b8-ec23182fd21d\") " pod="openshift-infra/auto-csr-approver-29553948-6trm4" Mar 11 13:48:00 crc kubenswrapper[4923]: I0311 13:48:00.415466 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bj4z\" (UniqueName: \"kubernetes.io/projected/8d9038fe-7cb9-4093-87b8-ec23182fd21d-kube-api-access-7bj4z\") pod \"auto-csr-approver-29553948-6trm4\" (UID: \"8d9038fe-7cb9-4093-87b8-ec23182fd21d\") " pod="openshift-infra/auto-csr-approver-29553948-6trm4" Mar 11 13:48:00 crc kubenswrapper[4923]: I0311 13:48:00.491295 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553948-6trm4" Mar 11 13:48:00 crc kubenswrapper[4923]: I0311 13:48:00.966077 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553948-6trm4"] Mar 11 13:48:00 crc kubenswrapper[4923]: W0311 13:48:00.984584 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8d9038fe_7cb9_4093_87b8_ec23182fd21d.slice/crio-7e248bcb0752eccb685870852344d52253f4fe01a0a83001a77cd1b29902882f WatchSource:0}: Error finding container 7e248bcb0752eccb685870852344d52253f4fe01a0a83001a77cd1b29902882f: Status 404 returned error can't find the container with id 7e248bcb0752eccb685870852344d52253f4fe01a0a83001a77cd1b29902882f Mar 11 13:48:00 crc kubenswrapper[4923]: I0311 13:48:00.986801 4923 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 11 13:48:01 crc kubenswrapper[4923]: I0311 13:48:01.515188 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553948-6trm4" event={"ID":"8d9038fe-7cb9-4093-87b8-ec23182fd21d","Type":"ContainerStarted","Data":"7e248bcb0752eccb685870852344d52253f4fe01a0a83001a77cd1b29902882f"} Mar 11 13:48:02 crc kubenswrapper[4923]: I0311 13:48:02.524729 4923 generic.go:334] "Generic (PLEG): container finished" podID="8d9038fe-7cb9-4093-87b8-ec23182fd21d" containerID="e63e7bfa4071508ae0e668c136170912ddd82ab4bf41a4ffb059078c851f4525" exitCode=0 Mar 11 13:48:02 crc kubenswrapper[4923]: I0311 13:48:02.524960 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553948-6trm4" event={"ID":"8d9038fe-7cb9-4093-87b8-ec23182fd21d","Type":"ContainerDied","Data":"e63e7bfa4071508ae0e668c136170912ddd82ab4bf41a4ffb059078c851f4525"} Mar 11 13:48:03 crc kubenswrapper[4923]: I0311 13:48:03.894781 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553948-6trm4" Mar 11 13:48:04 crc kubenswrapper[4923]: I0311 13:48:04.109490 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bj4z\" (UniqueName: \"kubernetes.io/projected/8d9038fe-7cb9-4093-87b8-ec23182fd21d-kube-api-access-7bj4z\") pod \"8d9038fe-7cb9-4093-87b8-ec23182fd21d\" (UID: \"8d9038fe-7cb9-4093-87b8-ec23182fd21d\") " Mar 11 13:48:04 crc kubenswrapper[4923]: I0311 13:48:04.116010 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d9038fe-7cb9-4093-87b8-ec23182fd21d-kube-api-access-7bj4z" (OuterVolumeSpecName: "kube-api-access-7bj4z") pod "8d9038fe-7cb9-4093-87b8-ec23182fd21d" (UID: "8d9038fe-7cb9-4093-87b8-ec23182fd21d"). InnerVolumeSpecName "kube-api-access-7bj4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:48:04 crc kubenswrapper[4923]: I0311 13:48:04.210858 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bj4z\" (UniqueName: \"kubernetes.io/projected/8d9038fe-7cb9-4093-87b8-ec23182fd21d-kube-api-access-7bj4z\") on node \"crc\" DevicePath \"\"" Mar 11 13:48:04 crc kubenswrapper[4923]: I0311 13:48:04.543020 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553948-6trm4" event={"ID":"8d9038fe-7cb9-4093-87b8-ec23182fd21d","Type":"ContainerDied","Data":"7e248bcb0752eccb685870852344d52253f4fe01a0a83001a77cd1b29902882f"} Mar 11 13:48:04 crc kubenswrapper[4923]: I0311 13:48:04.543080 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e248bcb0752eccb685870852344d52253f4fe01a0a83001a77cd1b29902882f" Mar 11 13:48:04 crc kubenswrapper[4923]: I0311 13:48:04.543110 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553948-6trm4" Mar 11 13:48:04 crc kubenswrapper[4923]: I0311 13:48:04.965171 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29553942-46xs7"] Mar 11 13:48:04 crc kubenswrapper[4923]: I0311 13:48:04.971690 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29553942-46xs7"] Mar 11 13:48:05 crc kubenswrapper[4923]: I0311 13:48:05.035150 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06883e63-9dba-48dd-a356-d990b50293ff" path="/var/lib/kubelet/pods/06883e63-9dba-48dd-a356-d990b50293ff/volumes" Mar 11 13:48:11 crc kubenswrapper[4923]: I0311 13:48:11.932195 4923 patch_prober.go:28] interesting pod/machine-config-daemon-fsfnw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 11 13:48:11 crc kubenswrapper[4923]: I0311 13:48:11.932728 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 11 13:48:34 crc kubenswrapper[4923]: I0311 13:48:34.612126 4923 scope.go:117] "RemoveContainer" containerID="b5f0c27328a332000af4dd0ed02cbf39e42d2d9095fd9b820964ede98e48c266" Mar 11 13:48:41 crc kubenswrapper[4923]: I0311 13:48:41.932172 4923 patch_prober.go:28] interesting pod/machine-config-daemon-fsfnw container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 11 13:48:41 crc kubenswrapper[4923]: I0311 13:48:41.932573 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 11 13:48:41 crc kubenswrapper[4923]: I0311 13:48:41.932627 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" Mar 11 13:48:41 crc kubenswrapper[4923]: I0311 13:48:41.933266 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c22d0f7656c6b989475681637e5cd2d3b3a534ac5fa3c03b27449f2838f08907"} pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 11 13:48:41 crc kubenswrapper[4923]: I0311 13:48:41.933372 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" podUID="356de301-ded3-4112-8406-1a28e44c2082" containerName="machine-config-daemon" containerID="cri-o://c22d0f7656c6b989475681637e5cd2d3b3a534ac5fa3c03b27449f2838f08907" gracePeriod=600 Mar 11 13:48:42 crc kubenswrapper[4923]: I0311 13:48:42.877694 4923 generic.go:334] "Generic (PLEG): container finished" podID="356de301-ded3-4112-8406-1a28e44c2082" containerID="c22d0f7656c6b989475681637e5cd2d3b3a534ac5fa3c03b27449f2838f08907" exitCode=0 Mar 11 13:48:42 crc kubenswrapper[4923]: I0311 13:48:42.877755 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" event={"ID":"356de301-ded3-4112-8406-1a28e44c2082","Type":"ContainerDied","Data":"c22d0f7656c6b989475681637e5cd2d3b3a534ac5fa3c03b27449f2838f08907"} Mar 11 13:48:42 crc kubenswrapper[4923]: I0311 13:48:42.878545 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-fsfnw" event={"ID":"356de301-ded3-4112-8406-1a28e44c2082","Type":"ContainerStarted","Data":"30447fbefbf3facc5602b20956dd27e6153364a61d0ef72914ca0d1c878ce38a"} Mar 11 13:48:42 crc kubenswrapper[4923]: I0311 13:48:42.878628 4923 scope.go:117] "RemoveContainer" containerID="3d359e768a3b56caad3d49994ffb4f7bcb69d87648b877548ba08cfe716e455c" Mar 11 13:48:52 crc kubenswrapper[4923]: I0311 13:48:52.039168 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-db-create-clm8k"] Mar 11 13:48:52 crc kubenswrapper[4923]: I0311 13:48:52.045740 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-db-create-clm8k"] Mar 11 13:48:53 crc kubenswrapper[4923]: I0311 13:48:53.039940 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60b9b9dd-cd32-48bb-9995-36bf024f9087" path="/var/lib/kubelet/pods/60b9b9dd-cd32-48bb-9995-36bf024f9087/volumes" Mar 11 13:48:53 crc kubenswrapper[4923]: I0311 13:48:53.043070 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-f524-account-create-update-cdc7k"] Mar 11 13:48:53 crc kubenswrapper[4923]: I0311 13:48:53.053213 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-f524-account-create-update-cdc7k"] Mar 11 13:48:55 crc kubenswrapper[4923]: I0311 13:48:55.039800 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19fc67d1-7c03-4a55-8baa-362e5a8a2344" path="/var/lib/kubelet/pods/19fc67d1-7c03-4a55-8baa-362e5a8a2344/volumes" Mar 11 13:49:13 crc kubenswrapper[4923]: I0311 13:49:13.053144 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-cjv6x"] Mar 11 13:49:13 crc kubenswrapper[4923]: I0311 13:49:13.057534 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-db-sync-cjv6x"] Mar 11 13:49:15 crc kubenswrapper[4923]: I0311 13:49:15.042732 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eae8474f-7523-4ee9-8f02-7be328688043" path="/var/lib/kubelet/pods/eae8474f-7523-4ee9-8f02-7be328688043/volumes" Mar 11 13:49:19 crc kubenswrapper[4923]: I0311 13:49:19.040109 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-nk5mc"] Mar 11 13:49:19 crc kubenswrapper[4923]: I0311 13:49:19.049254 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["cinder-kuttl-tests/keystone-bootstrap-nk5mc"] Mar 11 13:49:21 crc kubenswrapper[4923]: I0311 13:49:21.036746 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ad98c9d-3330-47ad-bf1e-a208ad0401ce" path="/var/lib/kubelet/pods/7ad98c9d-3330-47ad-bf1e-a208ad0401ce/volumes" Mar 11 13:49:34 crc kubenswrapper[4923]: I0311 13:49:34.719515 4923 scope.go:117] "RemoveContainer" containerID="1beb9859963659abd427eb2fe12c392da85f99762e7dcdccd8a85f9693702cd8" Mar 11 13:49:34 crc kubenswrapper[4923]: I0311 13:49:34.816816 4923 scope.go:117] "RemoveContainer" containerID="0848cd26093c6a84dbeea86a0c14b063883ca2130c21912db2187eba4e273c4b" Mar 11 13:49:34 crc kubenswrapper[4923]: I0311 13:49:34.842290 4923 scope.go:117] "RemoveContainer" containerID="d647de38b479270292e82d6d4c55e47b8449dd4756f683dc64630857abdfead1" Mar 11 13:49:34 crc kubenswrapper[4923]: I0311 13:49:34.871688 4923 scope.go:117] "RemoveContainer" containerID="acb527396cb316e35fa76097e8f1dbeda621bb250e35ce6a8055fd6164931b26" Mar 11 13:49:34 crc kubenswrapper[4923]: I0311 13:49:34.901151 4923 scope.go:117] "RemoveContainer" containerID="a7d1a0e725782476084f70d25781444849a26204d786d6bc8c3e9cea9397ce6c" Mar 11 13:49:34 crc kubenswrapper[4923]: I0311 13:49:34.926727 4923 scope.go:117] "RemoveContainer" containerID="4824ba162fa478a80e5086b293c197316a3e376385333aa14602521ad8e30f25" Mar 11 13:49:34 crc kubenswrapper[4923]: I0311 13:49:34.944526 4923 scope.go:117] "RemoveContainer" containerID="45b6ed22ba4610b78ac8d31fa24b882d84c2382b59d7e2e139798e7b1ec02dfa" Mar 11 13:49:34 crc kubenswrapper[4923]: I0311 13:49:34.977092 4923 scope.go:117] "RemoveContainer" containerID="3a11bd67548b38812031446c2b46e57b76c585f3ad01bfe43de580389986f5a9" Mar 11 13:50:00 crc kubenswrapper[4923]: I0311 13:50:00.145887 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29553950-8k65n"] Mar 11 13:50:00 crc kubenswrapper[4923]: E0311 13:50:00.148613 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d9038fe-7cb9-4093-87b8-ec23182fd21d" containerName="oc" Mar 11 13:50:00 crc kubenswrapper[4923]: I0311 13:50:00.148731 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d9038fe-7cb9-4093-87b8-ec23182fd21d" containerName="oc" Mar 11 13:50:00 crc kubenswrapper[4923]: I0311 13:50:00.148947 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d9038fe-7cb9-4093-87b8-ec23182fd21d" containerName="oc" Mar 11 13:50:00 crc kubenswrapper[4923]: I0311 13:50:00.149590 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553950-8k65n" Mar 11 13:50:00 crc kubenswrapper[4923]: I0311 13:50:00.155063 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553950-8k65n"] Mar 11 13:50:00 crc kubenswrapper[4923]: I0311 13:50:00.186898 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 11 13:50:00 crc kubenswrapper[4923]: I0311 13:50:00.186923 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 11 13:50:00 crc kubenswrapper[4923]: I0311 13:50:00.186982 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-nxck7" Mar 11 13:50:00 crc kubenswrapper[4923]: I0311 13:50:00.243705 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsg9x\" (UniqueName: \"kubernetes.io/projected/8c707ca0-5149-4069-9014-7a603eab87a7-kube-api-access-zsg9x\") pod \"auto-csr-approver-29553950-8k65n\" (UID: \"8c707ca0-5149-4069-9014-7a603eab87a7\") " pod="openshift-infra/auto-csr-approver-29553950-8k65n" Mar 11 13:50:00 crc kubenswrapper[4923]: I0311 13:50:00.344990 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsg9x\" (UniqueName: \"kubernetes.io/projected/8c707ca0-5149-4069-9014-7a603eab87a7-kube-api-access-zsg9x\") pod \"auto-csr-approver-29553950-8k65n\" (UID: \"8c707ca0-5149-4069-9014-7a603eab87a7\") " pod="openshift-infra/auto-csr-approver-29553950-8k65n" Mar 11 13:50:00 crc kubenswrapper[4923]: I0311 13:50:00.383332 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsg9x\" (UniqueName: \"kubernetes.io/projected/8c707ca0-5149-4069-9014-7a603eab87a7-kube-api-access-zsg9x\") pod \"auto-csr-approver-29553950-8k65n\" (UID: \"8c707ca0-5149-4069-9014-7a603eab87a7\") " pod="openshift-infra/auto-csr-approver-29553950-8k65n" Mar 11 13:50:00 crc kubenswrapper[4923]: I0311 13:50:00.510869 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553950-8k65n" Mar 11 13:50:00 crc kubenswrapper[4923]: I0311 13:50:00.980307 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29553950-8k65n"] Mar 11 13:50:00 crc kubenswrapper[4923]: W0311 13:50:00.990654 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8c707ca0_5149_4069_9014_7a603eab87a7.slice/crio-c185d6b05ecdeefe8770de537e782016ec39b0b24de8b383cce3902381f9ccf7 WatchSource:0}: Error finding container c185d6b05ecdeefe8770de537e782016ec39b0b24de8b383cce3902381f9ccf7: Status 404 returned error can't find the container with id c185d6b05ecdeefe8770de537e782016ec39b0b24de8b383cce3902381f9ccf7 Mar 11 13:50:01 crc kubenswrapper[4923]: I0311 13:50:01.571756 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553950-8k65n" event={"ID":"8c707ca0-5149-4069-9014-7a603eab87a7","Type":"ContainerStarted","Data":"c185d6b05ecdeefe8770de537e782016ec39b0b24de8b383cce3902381f9ccf7"} Mar 11 13:50:03 crc kubenswrapper[4923]: I0311 13:50:03.591441 4923 generic.go:334] "Generic (PLEG): container finished" podID="8c707ca0-5149-4069-9014-7a603eab87a7" containerID="e39a66883c3872e1795c02a361d84fc317b727d9aff67cbb73fc6deed3c24e19" exitCode=0 Mar 11 13:50:03 crc kubenswrapper[4923]: I0311 13:50:03.591533 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553950-8k65n" event={"ID":"8c707ca0-5149-4069-9014-7a603eab87a7","Type":"ContainerDied","Data":"e39a66883c3872e1795c02a361d84fc317b727d9aff67cbb73fc6deed3c24e19"} Mar 11 13:50:04 crc kubenswrapper[4923]: I0311 13:50:04.817100 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rn4lp"] Mar 11 13:50:04 crc kubenswrapper[4923]: I0311 13:50:04.822642 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rn4lp" Mar 11 13:50:04 crc kubenswrapper[4923]: I0311 13:50:04.833655 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rn4lp"] Mar 11 13:50:04 crc kubenswrapper[4923]: I0311 13:50:04.929820 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2-utilities\") pod \"redhat-operators-rn4lp\" (UID: \"34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2\") " pod="openshift-marketplace/redhat-operators-rn4lp" Mar 11 13:50:04 crc kubenswrapper[4923]: I0311 13:50:04.929919 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7jrr\" (UniqueName: \"kubernetes.io/projected/34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2-kube-api-access-t7jrr\") pod \"redhat-operators-rn4lp\" (UID: \"34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2\") " pod="openshift-marketplace/redhat-operators-rn4lp" Mar 11 13:50:04 crc kubenswrapper[4923]: I0311 13:50:04.929951 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2-catalog-content\") pod \"redhat-operators-rn4lp\" (UID: \"34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2\") " pod="openshift-marketplace/redhat-operators-rn4lp" Mar 11 13:50:04 crc kubenswrapper[4923]: I0311 13:50:04.944032 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553950-8k65n" Mar 11 13:50:05 crc kubenswrapper[4923]: I0311 13:50:05.030464 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsg9x\" (UniqueName: \"kubernetes.io/projected/8c707ca0-5149-4069-9014-7a603eab87a7-kube-api-access-zsg9x\") pod \"8c707ca0-5149-4069-9014-7a603eab87a7\" (UID: \"8c707ca0-5149-4069-9014-7a603eab87a7\") " Mar 11 13:50:05 crc kubenswrapper[4923]: I0311 13:50:05.030692 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2-utilities\") pod \"redhat-operators-rn4lp\" (UID: \"34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2\") " pod="openshift-marketplace/redhat-operators-rn4lp" Mar 11 13:50:05 crc kubenswrapper[4923]: I0311 13:50:05.030768 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7jrr\" (UniqueName: \"kubernetes.io/projected/34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2-kube-api-access-t7jrr\") pod \"redhat-operators-rn4lp\" (UID: \"34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2\") " pod="openshift-marketplace/redhat-operators-rn4lp" Mar 11 13:50:05 crc kubenswrapper[4923]: I0311 13:50:05.030790 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2-catalog-content\") pod \"redhat-operators-rn4lp\" (UID: \"34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2\") " pod="openshift-marketplace/redhat-operators-rn4lp" Mar 11 13:50:05 crc kubenswrapper[4923]: I0311 13:50:05.031168 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2-utilities\") pod \"redhat-operators-rn4lp\" (UID: \"34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2\") " pod="openshift-marketplace/redhat-operators-rn4lp" Mar 11 13:50:05 crc kubenswrapper[4923]: I0311 13:50:05.031230 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2-catalog-content\") pod \"redhat-operators-rn4lp\" (UID: \"34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2\") " pod="openshift-marketplace/redhat-operators-rn4lp" Mar 11 13:50:05 crc kubenswrapper[4923]: I0311 13:50:05.037292 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c707ca0-5149-4069-9014-7a603eab87a7-kube-api-access-zsg9x" (OuterVolumeSpecName: "kube-api-access-zsg9x") pod "8c707ca0-5149-4069-9014-7a603eab87a7" (UID: "8c707ca0-5149-4069-9014-7a603eab87a7"). InnerVolumeSpecName "kube-api-access-zsg9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:50:05 crc kubenswrapper[4923]: I0311 13:50:05.047604 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7jrr\" (UniqueName: \"kubernetes.io/projected/34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2-kube-api-access-t7jrr\") pod \"redhat-operators-rn4lp\" (UID: \"34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2\") " pod="openshift-marketplace/redhat-operators-rn4lp" Mar 11 13:50:05 crc kubenswrapper[4923]: I0311 13:50:05.132605 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsg9x\" (UniqueName: \"kubernetes.io/projected/8c707ca0-5149-4069-9014-7a603eab87a7-kube-api-access-zsg9x\") on node \"crc\" DevicePath \"\"" Mar 11 13:50:05 crc kubenswrapper[4923]: I0311 13:50:05.151576 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rn4lp" Mar 11 13:50:05 crc kubenswrapper[4923]: I0311 13:50:05.539738 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rn4lp"] Mar 11 13:50:05 crc kubenswrapper[4923]: W0311 13:50:05.543294 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34f54739_0e2b_4c08_ba32_fe1ddb5fb3a2.slice/crio-4747c045d797270d1758b1560e5d4d789596f87ce16b052c635f5db08c8d7385 WatchSource:0}: Error finding container 4747c045d797270d1758b1560e5d4d789596f87ce16b052c635f5db08c8d7385: Status 404 returned error can't find the container with id 4747c045d797270d1758b1560e5d4d789596f87ce16b052c635f5db08c8d7385 Mar 11 13:50:05 crc kubenswrapper[4923]: I0311 13:50:05.608361 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rn4lp" event={"ID":"34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2","Type":"ContainerStarted","Data":"4747c045d797270d1758b1560e5d4d789596f87ce16b052c635f5db08c8d7385"} Mar 11 13:50:05 crc kubenswrapper[4923]: I0311 13:50:05.611169 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29553950-8k65n" event={"ID":"8c707ca0-5149-4069-9014-7a603eab87a7","Type":"ContainerDied","Data":"c185d6b05ecdeefe8770de537e782016ec39b0b24de8b383cce3902381f9ccf7"} Mar 11 13:50:05 crc kubenswrapper[4923]: I0311 13:50:05.611199 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c185d6b05ecdeefe8770de537e782016ec39b0b24de8b383cce3902381f9ccf7" Mar 11 13:50:05 crc kubenswrapper[4923]: I0311 13:50:05.611249 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29553950-8k65n" Mar 11 13:50:06 crc kubenswrapper[4923]: I0311 13:50:06.014562 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29553944-hktnj"] Mar 11 13:50:06 crc kubenswrapper[4923]: I0311 13:50:06.020016 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29553944-hktnj"] Mar 11 13:50:06 crc kubenswrapper[4923]: I0311 13:50:06.623182 4923 generic.go:334] "Generic (PLEG): container finished" podID="34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2" containerID="5c91ff7d33b119298490940482d27bd5fd8f38e14fdd237a6f9e714592123257" exitCode=0 Mar 11 13:50:06 crc kubenswrapper[4923]: I0311 13:50:06.623436 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rn4lp" event={"ID":"34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2","Type":"ContainerDied","Data":"5c91ff7d33b119298490940482d27bd5fd8f38e14fdd237a6f9e714592123257"} Mar 11 13:50:07 crc kubenswrapper[4923]: I0311 13:50:07.040218 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e956abd-be22-488a-895d-171f5ebe495b" path="/var/lib/kubelet/pods/9e956abd-be22-488a-895d-171f5ebe495b/volumes" Mar 11 13:50:07 crc kubenswrapper[4923]: I0311 13:50:07.631765 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rn4lp" event={"ID":"34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2","Type":"ContainerStarted","Data":"dde391eadf7bcfacad606c95f69b392e863d9c92cf40b4551cc578f9f7265e38"} Mar 11 13:50:08 crc kubenswrapper[4923]: I0311 13:50:08.647585 4923 generic.go:334] "Generic (PLEG): container finished" podID="34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2" containerID="dde391eadf7bcfacad606c95f69b392e863d9c92cf40b4551cc578f9f7265e38" exitCode=0 Mar 11 13:50:08 crc kubenswrapper[4923]: I0311 13:50:08.647697 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rn4lp" event={"ID":"34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2","Type":"ContainerDied","Data":"dde391eadf7bcfacad606c95f69b392e863d9c92cf40b4551cc578f9f7265e38"} Mar 11 13:50:09 crc kubenswrapper[4923]: I0311 13:50:09.664249 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rn4lp" event={"ID":"34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2","Type":"ContainerStarted","Data":"505b1c7c2a634fd41440cd804eddaa4df1272228bd1b05c4c1c360bdfec15b49"} Mar 11 13:50:09 crc kubenswrapper[4923]: I0311 13:50:09.696793 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rn4lp" podStartSLOduration=3.222482406 podStartE2EDuration="5.696759664s" podCreationTimestamp="2026-03-11 13:50:04 +0000 UTC" firstStartedPulling="2026-03-11 13:50:06.626922649 +0000 UTC m=+1654.308824393" lastFinishedPulling="2026-03-11 13:50:09.101199937 +0000 UTC m=+1656.783101651" observedRunningTime="2026-03-11 13:50:09.695663665 +0000 UTC m=+1657.377565419" watchObservedRunningTime="2026-03-11 13:50:09.696759664 +0000 UTC m=+1657.378661408" Mar 11 13:50:15 crc kubenswrapper[4923]: I0311 13:50:15.152571 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rn4lp" Mar 11 13:50:15 crc kubenswrapper[4923]: I0311 13:50:15.153672 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rn4lp" Mar 11 13:50:16 crc kubenswrapper[4923]: I0311 13:50:16.217198 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rn4lp" podUID="34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2" containerName="registry-server" probeResult="failure" output=< Mar 11 13:50:16 crc kubenswrapper[4923]: timeout: failed to connect service ":50051" within 1s Mar 11 13:50:16 crc kubenswrapper[4923]: > Mar 11 13:50:25 crc kubenswrapper[4923]: I0311 13:50:25.205923 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rn4lp" Mar 11 13:50:25 crc kubenswrapper[4923]: I0311 13:50:25.281780 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rn4lp" Mar 11 13:50:25 crc kubenswrapper[4923]: I0311 13:50:25.451723 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rn4lp"] Mar 11 13:50:26 crc kubenswrapper[4923]: I0311 13:50:26.817765 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rn4lp" podUID="34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2" containerName="registry-server" containerID="cri-o://505b1c7c2a634fd41440cd804eddaa4df1272228bd1b05c4c1c360bdfec15b49" gracePeriod=2 Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.162638 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rn4lp" Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.340015 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7jrr\" (UniqueName: \"kubernetes.io/projected/34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2-kube-api-access-t7jrr\") pod \"34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2\" (UID: \"34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2\") " Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.340119 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2-catalog-content\") pod \"34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2\" (UID: \"34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2\") " Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.340165 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2-utilities\") pod \"34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2\" (UID: \"34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2\") " Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.341303 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2-utilities" (OuterVolumeSpecName: "utilities") pod "34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2" (UID: "34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.347946 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2-kube-api-access-t7jrr" (OuterVolumeSpecName: "kube-api-access-t7jrr") pod "34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2" (UID: "34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2"). InnerVolumeSpecName "kube-api-access-t7jrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.441581 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7jrr\" (UniqueName: \"kubernetes.io/projected/34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2-kube-api-access-t7jrr\") on node \"crc\" DevicePath \"\"" Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.441612 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2-utilities\") on node \"crc\" DevicePath \"\"" Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.518717 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2" (UID: "34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.543136 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.830624 4923 generic.go:334] "Generic (PLEG): container finished" podID="34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2" containerID="505b1c7c2a634fd41440cd804eddaa4df1272228bd1b05c4c1c360bdfec15b49" exitCode=0 Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.830677 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rn4lp" event={"ID":"34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2","Type":"ContainerDied","Data":"505b1c7c2a634fd41440cd804eddaa4df1272228bd1b05c4c1c360bdfec15b49"} Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.830734 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rn4lp" event={"ID":"34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2","Type":"ContainerDied","Data":"4747c045d797270d1758b1560e5d4d789596f87ce16b052c635f5db08c8d7385"} Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.830744 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rn4lp" Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.830794 4923 scope.go:117] "RemoveContainer" containerID="505b1c7c2a634fd41440cd804eddaa4df1272228bd1b05c4c1c360bdfec15b49" Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.854431 4923 scope.go:117] "RemoveContainer" containerID="dde391eadf7bcfacad606c95f69b392e863d9c92cf40b4551cc578f9f7265e38" Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.896456 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rn4lp"] Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.899765 4923 scope.go:117] "RemoveContainer" containerID="5c91ff7d33b119298490940482d27bd5fd8f38e14fdd237a6f9e714592123257" Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.906812 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rn4lp"] Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.928740 4923 scope.go:117] "RemoveContainer" containerID="505b1c7c2a634fd41440cd804eddaa4df1272228bd1b05c4c1c360bdfec15b49" Mar 11 13:50:27 crc kubenswrapper[4923]: E0311 13:50:27.929283 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"505b1c7c2a634fd41440cd804eddaa4df1272228bd1b05c4c1c360bdfec15b49\": container with ID starting with 505b1c7c2a634fd41440cd804eddaa4df1272228bd1b05c4c1c360bdfec15b49 not found: ID does not exist" containerID="505b1c7c2a634fd41440cd804eddaa4df1272228bd1b05c4c1c360bdfec15b49" Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.929328 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"505b1c7c2a634fd41440cd804eddaa4df1272228bd1b05c4c1c360bdfec15b49"} err="failed to get container status \"505b1c7c2a634fd41440cd804eddaa4df1272228bd1b05c4c1c360bdfec15b49\": rpc error: code = NotFound desc = could not find container \"505b1c7c2a634fd41440cd804eddaa4df1272228bd1b05c4c1c360bdfec15b49\": container with ID starting with 505b1c7c2a634fd41440cd804eddaa4df1272228bd1b05c4c1c360bdfec15b49 not found: ID does not exist" Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.929374 4923 scope.go:117] "RemoveContainer" containerID="dde391eadf7bcfacad606c95f69b392e863d9c92cf40b4551cc578f9f7265e38" Mar 11 13:50:27 crc kubenswrapper[4923]: E0311 13:50:27.930668 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dde391eadf7bcfacad606c95f69b392e863d9c92cf40b4551cc578f9f7265e38\": container with ID starting with dde391eadf7bcfacad606c95f69b392e863d9c92cf40b4551cc578f9f7265e38 not found: ID does not exist" containerID="dde391eadf7bcfacad606c95f69b392e863d9c92cf40b4551cc578f9f7265e38" Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.930700 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dde391eadf7bcfacad606c95f69b392e863d9c92cf40b4551cc578f9f7265e38"} err="failed to get container status \"dde391eadf7bcfacad606c95f69b392e863d9c92cf40b4551cc578f9f7265e38\": rpc error: code = NotFound desc = could not find container \"dde391eadf7bcfacad606c95f69b392e863d9c92cf40b4551cc578f9f7265e38\": container with ID starting with dde391eadf7bcfacad606c95f69b392e863d9c92cf40b4551cc578f9f7265e38 not found: ID does not exist" Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.930718 4923 scope.go:117] "RemoveContainer" containerID="5c91ff7d33b119298490940482d27bd5fd8f38e14fdd237a6f9e714592123257" Mar 11 13:50:27 crc kubenswrapper[4923]: E0311 13:50:27.930988 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c91ff7d33b119298490940482d27bd5fd8f38e14fdd237a6f9e714592123257\": container with ID starting with 5c91ff7d33b119298490940482d27bd5fd8f38e14fdd237a6f9e714592123257 not found: ID does not exist" containerID="5c91ff7d33b119298490940482d27bd5fd8f38e14fdd237a6f9e714592123257" Mar 11 13:50:27 crc kubenswrapper[4923]: I0311 13:50:27.931015 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c91ff7d33b119298490940482d27bd5fd8f38e14fdd237a6f9e714592123257"} err="failed to get container status \"5c91ff7d33b119298490940482d27bd5fd8f38e14fdd237a6f9e714592123257\": rpc error: code = NotFound desc = could not find container \"5c91ff7d33b119298490940482d27bd5fd8f38e14fdd237a6f9e714592123257\": container with ID starting with 5c91ff7d33b119298490940482d27bd5fd8f38e14fdd237a6f9e714592123257 not found: ID does not exist" Mar 11 13:50:29 crc kubenswrapper[4923]: I0311 13:50:29.038314 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2" path="/var/lib/kubelet/pods/34f54739-0e2b-4c08-ba32-fe1ddb5fb3a2/volumes" Mar 11 13:50:35 crc kubenswrapper[4923]: I0311 13:50:35.123960 4923 scope.go:117] "RemoveContainer" containerID="4f86578704f9bdee6ff6aa583785d75e6d018c9cb4f8cf5cd6f59c34460769ce" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515154271530024450 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015154271531017366 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015154265740016516 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015154265740015466 5ustar corecore